key: cord-0280452-rr5axdwn authors: Roncalli, Thierry; Karray-Meziou, Fatma; Pan, Franccois; Regnault, Margaux title: Liquidity Stress Testing in Asset Management -- Part 1. Modeling the Liability Liquidity Risk date: 2021-01-06 journal: nan DOI: 10.13140/rg.2.2.14893.72165 sha: beacabc816c9501a042e84cdad25444478ae7e58 doc_id: 280452 cord_uid: rr5axdwn This article is part of a comprehensive research project on liquidity risk in asset management, which can be divided into three dimensions. The first dimension covers liability liquidity risk (or funding liquidity) modeling, the second dimension focuses on asset liquidity risk (or market liquidity) modeling, and the third dimension considers asset-liability liquidity risk management (or asset-liability matching). The purpose of this research is to propose a methodological and practical framework in order to perform liquidity stress testing programs, which comply with regulatory guidelines (ESMA, 2019) and are useful for fund managers. The review of the academic literature and professional research studies shows that there is a lack of standardized and analytical models. The aim of this research project is then to fill the gap with the goal to develop mathematical and statistical approaches, and provide appropriate answers. In this first part that focuses on liability liquidity risk modeling, we propose several statistical models for estimating redemption shocks. The historical approach must be complemented by an analytical approach based on zero-inflated models if we want to understand the true parameters that influence the redemption shocks. Moreover, we must also distinguish aggregate population models and individual-based models if we want to develop behavioral approaches. Once these different statistical models are calibrated, the second big issue is the risk measure to assess normal and stressed redemption shocks. Finally, the last issue is to develop a factor model that can translate stress scenarios on market risk factors into stress scenarios on fund liabilities. Liquidity stress testing in asset management is a complex topic because it is related to three dimensions -liquidity risk, asset management and stress testing, whose linkages have been little studied and are hard to capture. First, liquidity is certainly the risk that is the most difficult to model with the systemic risk. If we consider market, credit, operational and counterparty risks, there is a huge amount of academic literature on these topics in terms of models, statistical inference and analysis. In terms of liquidity risk, the number of practical studies and applied approaches is limited. Even though a great deal of research has been completed on this subject, much of it is overly focused on descriptive analyses of liquidity, or its impact on systemic risk, or policy rules for financial regulation. Moreover, this research generally focuses on the banking sector (Grillet-Aubert, 2018) . For instance, the liquidity coverage ratio (LCR) and the net stable funding ratio (NSFR) of the Basel III regulatory framework are of no help when measuring the liquidity risk in asset management. In fact, the concept of liquidity risk in asset management is not well defined. More generally, it is a recent subject and we must admit that the tools and models used in asset management are very much lagging those developed in the banking sector. This is why the culture of asset-liability management (ALM) is poor in investment management, both on the side of asset managers and asset owners. Therefore, if we add the third dimension, stress testing, we obtain an unknown and obscure topic, because the combination of liquidity risk and stress testing applied to asset management is a new and difficult task. This is not the first time that the regulatory environment has sped up the development of a risk management framework. Previous occurrences include the case of market risk with the Amendment of the first Basel Accord (BCBS, 1996) , credit risk with the second consultative paper on Basel II (BCBS, 2001) , credit valuation adjustment with the publication of the Basel III Accord (BCBS, 2010) , interest rate risk in the banking book with the IR-RBB guidelines (BCBS, 2016) , etc. However, the measurement of these risks had already benefited from the existence of analytical models developed by academics and professionals. One exception was operational risk, since banks started from a blank page when asked to measure it (BCBS, 1999) . Asset managers now face a similar situation at this moment. Between 2015 and 2018, the US Securities and Exchange Commission established several rules governing liquidity management (SEC, 2015 (SEC, , 2016 (SEC, , 2018a . In particular, Rule 22e-4 requires investment funds to classify their positions in one of four liquidity buckets (highly liquid investments, moderately liquid investments, less liquid investments and illiquid investments), establish a liquid investment minimum, and develop policies and procedures on redemptions in kind. From September 2020, European asset managers must also comply with new guidelines on liquidity stress testing (LST) published by the European Securities and Markets Authority (ESMA, 2019). These different regulations are rooted in the agenda proposed by the Financial Stability Board to monitor and manage systemic risk of non-bank non-insurer systemically important financial institutions (FSB, 2010) . Even if the original works of the FSB were biased, the idea that the asset management industry can contribute to systemic risk has gained ground and is now widely accepted. Indeed, FSB (2015) confused systemic risk and systematic market risk (Roncalli and Weisang, 2015a) . However, Roncalli and Weisang (2015b) showed that "the liquidation channel is the main component of systemic risk to which the asset management industry contributes". In this context, liquidity is the major risk posed by the asset management industry that regulators must control. But liquidity risk is not only a concern for regulators. It must also be a priority for asset managers. The crisis of money market funds in the fourth quarter of 2008 demonstrated the fragility of some fund managers (Schmidt et al., 2016) . Market liquidity deteriorated in March and April 2020, triggering a liquidity shock on some investment funds and strategies. However, aside from the 2008 Global Financial Crisis and 2020 coronavirus pandemic, which have put all asset managers under pressure, the last ten years have demonstrated that liquidity is also an individual risk for fund managers. It was especially true during episodes of flash crash 1 , where fund managers reacted differently. In a similar way, idiosyncratic liquidity events may affect asset managers at the individual level (Thompson, 2019) . Following some high-profile fund suspensions in mid-2019, asset managers received requests from asset owners to describe their liquidity policies and conduct a liquidity review of their portfolios. Therefore, we notice that liquidity is increasingly becoming a priority for asset managers for three main reasons, because it is a reputational risk, they are challenged by asset owners and it can be a vulnerability factor for financial performance. However, even though liquidity stress testing in asset management has become one of the hot topics in finance, it has attracted few academics and professionals, implying that the research on this subject is not as dynamic as one might expect. In fact, it is at the same stage as operational risk was in the early 2000s, when there was no academic research on this topic. And it is also at the stage of ALM banking risk, where the most significant contributions have come from professionals. Since liquidity stress testing in asset management is an assetliability management exercise, modeling progress mainly comes from professionals, because the subject is so specific, requires business expertise and must be underpinned by industrylevel data. This is obviously an enormous hurdle for academics, and this explains the lack of modeling and scientific approach that asset managers encounter when they want to develop a liquidity stress testing framework. Therefore, the objective of this research is twofold. First, the idea is to provide a mathematical and statistical formalization to professionals in order to go beyond expert qualitative judgement. Second, the aim is to assist academics in understanding this topic. This is important, because academic research generally boosts the development of analytical models, which are essential for implementing liquidity stress testing programs in asset management. Liquidity stress testing in asset management involves so many dimensions that we have decided to split this research into three parts: 1. liability liquidity risk modeling; 2. asset liquidity risk modeling; 3 . asset-liability liquidity risk management. Indeed, managing liquidity risk consists of three steps. First, we have to model the liability liquidity of the investment fund, especially the redemption shocks. By construction, this step must incorporate client behavior. Second, we have to develop a liquidity model for assets. For that, we must specify a transaction cost model beyond the traditional bid-ask spread measure. In particular, the model must incorporate two dimensions: price impact and trading limits. These first two steps make the distinction between funding liquidity and market liquidity. As noticed by Brunnermeier and Pedersen (2009) , these two types of liquidity may be correlated. However, we suppose that they are independent at this level of analysis. While the first step gives the liquidity amount of the investment fund that can be required by the investors, the second step gives the liquidity amount of the investment fund that can be available in the market. Therefore, the third step corresponds to the assetliability management in terms of liquidity, that is the matching process between required liquidity and available liquidity. This implies defining the part of the redemption shock that can be managed by asset liquidation and the associated liquidity costs. It also implies defining the liquidity tools that can be put in place in order to manage the non-covered part of the redemption shock or the liquidation shortfall. For instance, a liquidity buffer is an example of one of these tools, but this is not the only solution. Redemption gates, side pockets and redemptions in kind are alternative methods, but they are extreme solutions that may break the fiduciary duties and liquidity promises of asset managers. Swing pricing is also an important ALM tool, and is a challenging question when we consider the fair calibration of swing prices. The three-stage process has many advantages in terms of modeling. First, it splits a complex question into three independent and more manageable problems. This is particularly the case of liability and asset modeling. Second, managing liquidity risk becomes a sequential process, where the starting point is clearly identified. As shown in Figure 1 , we should begin with the liability risk. Indeed, if we observe no inflows or outflows, the process stops here. As such, the first stage determines the amount to sell in the market and it is measured with respect to the investor behavior. The liquidity risk has its roots in the severity of the redemption shock. Market liquidity is part of the second phase. Depending on the redemption size and the liquidity of the market, the fund manager will decide the best solution to adopt. And the sequential process will conclude with the action of the fund manager 2 . Finally, the third advantage concerns the feasibility of stress testing programs. In this approach, stress testing concerns the two independent dimensions. We can stress the liquidity on the liability side, or we can stress the liquidity on the asset side, or both, but the rule is simple. In the sequential approach, the liability of the investment fund is the central node of the liquidity risk, and the vertex of the liquidity network. However, it is not so simple, because the three nodes can be interconnected ( Figure 2 ). If market liquidity deteriorates sharply, investors may be incited to redeem in order to avoid a liquidity trap. In this case, funding liquidity is impacted by market liquidity, reinforcing the feedback loop between funding and market liquidity, which is described by Brunnermeier and Pedersen (2009) . But this is not the only loop. For instance, the choice of an ALM decision may also influence funding liquidity. If one asset manager decides to suspend redemptions, it may be a signal for the investors of the other asset managers if they continue to meet redemptions. Again, we may observe a feedback loop between funding liquidity and asset-liability management. Finally, it is also obvious that market liquidity is related to ALM decisions, because of many factors such as trading policies, the first-mover advantage and crowding effects (Roncalli and Weisang, 2015a) . It follows that the liquidity risk given in Figure 1 is best described by the dense and fully connected network given in Figure 2 . Nevertheless, developing a statistical model that takes into account the three reinforcing loops is not straightforward and certainly too complex for professional use. Therefore, it is more realistic to adjust and update the sequential models with second-round effects than to have an integrated dynamic model. Liquidity is a long-standing issue and also an elusive concept. This is particularly true in asset management, where liquidity covers several interpretations. For example, some asset classes are considered as highly liquid whereas other asset classes are illiquid. In the first category, we generally find government bonds and large cap stocks. The last category includes real estate and private equities. However, categorizing liquidity of a security is not easy and there is no consensus. Let us consider for example Rule 22e-4(b) that is applied in the US. The proposed rule was based on the ability to convert the security to cash within a given period and distinguished six buckets: (a) convertible to cash within 1 business day, (b) convertible to cash within 2-3 business days, (c) convertible to cash within 4-7 calendar days, (d) convertible to cash within 8-15 calendar days, (e) convertible to cash within 16-30 calendar days (f) convertible to cash in more than one month. Finally, the adopted rule is the following: 1. highly liquid investments (convertible to cash within three business days); 2. moderately liquid investments (convertible to cash within four to seven calendar days); 3 . less liquid investments (expected to be sold in more than seven calendar days); 4 . illiquid investments (cannot be sold in seven calendar days or less without significant price impact). Classifying a security into a bucket may be different from one fund manager to another. Moreover, the previous categories depend on the market conditions. Nevertheless, even if the current market liquidity is abundant, securities that can be categorized in the first bucket must also face episodes of liquidity shortage (Blanqué and Mortier, 2019a) . A typical example concerns government bonds facing idiosyncratic risks. Blanqué and Mortier (2019a) gave the case of Italian bonds in 2018 during the discussion on the budget deficit. However, most of the time, when we consider the liquidity of an asset class, we assume that it is static. Certainly, this way of thinking reflects the practice of portfolio management. Indeed, it is common to include a constant illiquidity premium when estimating the expected returns of illiquid assets. But investors should stick to their investments without rebalancing and trading if they want to capture this illiquidity premium. The split between liquid and illiquid investments does not help, because it is related to the absolute level of asset illiquidity, and not liquidity dynamics. However, the issue is more complex: " [...] there is also broad belief among users of financial liquidity -traders, investors and central bankers -that the principal challenge is not the average level of financial liquidity. . . but its variability and uncertainty" (Persaud, 2003) . This observation is important because it is related to the liquidity question from a regulatory point of view. The liquidity risk of private equities or real assets is not a big concern for regulators, because one knows that these asset classes are illiquid. Even if they become more illiquid at some point, this should not dramatically influence investors (asset managers and owners). Regulators and investors are more concerned by securities that are liquid under some market conditions and illiquid under other market conditions. At first sight, it is therefore a paradox that liquidity stress testing programs must mainly focus on highly or moderately liquid instruments than on illiquid instruments. In fact, liquidity does not like surprises and changes. This is why the liquidity issue is related to the cross-section of the expected illiquidity premium for illiquid assets, but to the time-series illiquidity variance for liquid assets. This is all the more important that the liquidity risk must be measured and managed in a stress testing framework, which adds another layer of complexity. Indeed, stress scenarios are always difficult to interpret, and calibrating them is a balancing act, because they must correspond to extreme but also plausible events (Roncalli, 2020) . This is why the historical method is the most used approach when performing stress testing. However, it is very poor and not flexible in terms of risk management. Parametric approaches must be preferred since stress periods are very heterogenous and outcomes are uncertain. Therefore, it makes more sense to estimate and use stressed liquidity parameters than directly estimate a stressed liquidity outcome. In this approach, the normal model is the baseline model on which we could apply scenario analysis on the different parameters that define the liquidity model. This is certainly the best way to proceed if we want to develop a factor-based liquidity stress testing program, which is an important issue for fund management. Otherwise, liquidity stress testing would be likely to remain a regulatory constraint or a pure exercise of risk measurement, but certainly not a risk management process supporting investment policies and fund management. This paper is organized as follows. Section Two introduces the concept of redemption rates and defines the historical approach of liquidity stress testing. In Section Three, we consider parametric models that can be used to estimate redemption shocks. This implies making the distinction between the redemption event and the redemption amount. From a statistical point of view, this is equivalent to modeling the redemption frequency and the redemption severity. After having developed an aggregate population model, we consider an individual-based model. It can be considered as a first attempt to develop a behavioral model, which is the central theme of Section Four. We analyze the simple case where redemptions between investors are independent and then extend the model where redemptions are correlated to take into account spillover effects and contagion risk. Then, we develop factor-based models of liquidity stress testing in Section Five. Finally, Section Six offers some concluding remarks. In order to assess the liquidity risk of an investment fund, we must model its 'funding' liquidity. Therefore, managing the liquidity in asset management looks like a banking assetliability management process (Roncalli, 2020) . However, there is a major difference since banking ALM concerns both balance sheet and income statement. This is not the case of an investment fund, because we only focus on its balance sheet and the objective is to model the redemption flows. In order to define the liability risk, we first have to understand the balance sheet of a collective investment fund. A simplified illustration is given in Figure 3 for a mutual fund. The total (gross) assets A (t) correspond to the market value of the investment portfolio. They include stocks, bonds and all financial instruments that are invested. On the liability side, we have two main balance sheet items. The first one corresponds to the debits D (t), which are also called current or accrued liabilities. They are all the expenses incurred by the mutual fund. For instance, the current liabilities include money owed to lending banks, fees owed to the fund manager and the custodian, etc. The second liability item is the unit capital C (t), which is owned by the investors. Each investor owns a number of units (or shares) and is referred to as a 'unitholder '. This unit capital is equivalent to the equity concept of a financial institution. A unitholder is then also called a shareholder in reference to capital markets. The total net assets (TNA) equal the total value of assets less the current or accrued liabilities: The net asset value (NAV) represents the share price or the unit price. It is equal to: where the total number N (t) of shares or units in issue is the sum of all units owned by all unitholders. The previous accounting rules show that the capital is exactly equal to the total net assets, which is also called the assets under management (AUM). The investment fund's capital is therefore an endogenous variable and depends on the performance of the total net assets: At time t+1, we assume that the portfolio's return is equal to R (t + 1). Since D (t) A (t), it follows that: meaning that: The investment fund's capital is therefore time-varying. It increases when the performance of the asset is positive, and it decreases otherwise. Remark 1 In the sequel, we assume that the mutual fund is priced daily, meaning that the NAV of the mutual fund is calculated at the end of the market day. Therefore, the time t represents the current market day, whereas the time t + 1 corresponds to the next market day. Let us now introduce the impact of subscriptions and redemptions. In this case, new and current investors may purchase new mutual fund units, while existing investors may redeem all or part of their shares. Subscription and redemption orders must be known by the fund manager before t + 1 in order to be executed. In this case, the number of units becomes: where N + (t + 1) is the number of units to be created and N − (t + 1) is the number of units to be redeemed. At time t + 1, we have: We deduce that: where F + (t + 1) = N + (t + 1) · NAV (t + 1) and F − (t + 1) = N − (t + 1) · NAV (t + 1) are the investment inflows and outflows. Again, we notice that the investment fund's capital is time-varying and depends on the fund flows. From Equation (2), we deduce that: The current net assets are approximatively equal to the previous net assets plus the performance value plus the net flow. We retrieve the famous formula of Sirri and Tufano (1998) when we want to estimate the net flow from the NAV and TNA of the fund: Since the capital is a residual, we face three liability risks. The first one deals with the accrued liabilities D (t). Generally, the debits are a very small part of the liabilities. However, we can potentially face some situations where the debits are larger than the assets, implying that the net asset value becomes negative. In particular, this type of situation occurs when the fund is highly leveraged. The second risk concerns the inflows. If the investment fund has a big subscription, it may have some difficulties buying the financial instruments. For instance, this type of situation may occur when the fund must buy fixed-income securities in a bond bull market and it is difficult to find investors who are looking to sell bonds. The third liability risk is produced by the outflows. In this case, the fund manager must sell assets, which could be difficult in illiquid and stressed market conditions. The last two situations are produced when supply and demand dynamics are totally unbalanced (higher supply for buying assets or higher demand for selling assets). In this article, we focus on the third liability risk, which is also called redemption risk. In order to assess an investment fund's redemption risk, we need an objective measurement system, which is well scaled. For instance, the outflows F − (t) are not very interesting, because they depend on the investment fund's assets under management. In fact, they must be scaled in order to be a homogeneous measure that can be used to compare the redemption behavior across time, across funds and across investors. The (gross) redemption rate is defined as the ratio between the fund's redemption flows and total net assets: We verify the property that R (t) ∈ [0, 1]. For example, if we observe an outflow of $100 mn for a fund of $5 bn, we have R (t) = 100/5 000 = 2%. In the case where the outflow is $10 mn and the fund size is $100 mn, the redemption rate is equal to 10%. The redemption is more severe for the small fund than for the large fund. We notice that Equation (4) is used to calculate the ex-post redemption rate, meaning that the value of outflows is known. Therefore, Equation (4) corresponds to the definition of the redemption rate, but it can also be used to estimate or predict the redemption flows. Indeed, we have:F − (t + 1) = R (t + 1) · TNA (t) In this case, R (t + 1) is a random variable and is not known at the current time t. By assuming that redemption rates are stationary, the challenge is then to model the associated probability distribution F. The guidelines on the liquidity stress testing published by ESMA (2019) refer to both gross and net redemptions: "LST should be adapted appropriately to each fund, including by adapting: [...] the assumptions regarding investor behaviour (gross and net redemptions)" (ESMA, 2019, page 36). Following this remark, we can also define the net flow rate by considering both inflows and outflows: This quantity is more complex than the previous one, because it cannot be used from an ex-ante point of view: The reason is that the outflows are bounded and cannot exceed the current assets under management. This is not the case for the inflows. For example, we consider a fund with a size of $100 mn. By construction, we have 3F − (t + 1) ≤ 100, but we can imagine that F + (t + 1) > 100. The fund size can double or triple, in particular when the investment fund is young and small. Nevertheless, the use of net flows is not foolish since the true liability risk of the fund is on the net flows. If the fund manager faces a large redemption, which is offset by a large subscription, there is no liquidity risk. The issue is that the use of net flows is difficult to justify in stress periods. In these cases, inflows generally disappear and the probability distribution of R ± (t) may not reflect the liability risk in a stress testing framework. For example, let us consider an asset class that has experienced a bull market over the last three years. Certainly, we will mainly observe positive net flows and a very small number of observations with negative net flows. We may think that these data are not relevant for building stress scenarios. More generally, if an asset manager uses net flow rates for stress testing purposes, only the observations during historical stress periods are relevant, meaning that the calibration is based on a small fraction of the dataset. In fact, the use of net flows is motivated by other considerations. Indeed, the computation of R (t) requires us to know the outflows F − (t) exactly. Moreover, as we will see later, R (t) must be computed for all the investor categories that are present in the fund (retail, private banking, institutional, etc.) . This implies in-depth knowledge of the fund's balance sheet liability, meaning that the asset manager must have a database with all the flows of all the investors on a daily basis. From an industrial point of view, this is a big challenge in terms of IT systems between the asset manager and the custodian. This is why many asset managers don't have the disaggregated information on the liability flows. An alternative measure is to compute the net redemption rate, which corresponds to the negative part of the net flow rate: It has the good mathematical property that R − (t) ∈ [0, 1]. Indeed, we have: and its maximum value is reached when F − (t) = TNA (t) and F + (t) = 0. Moreover, we notice that the net redemption rate is equal to the gross redemption rate when there are no inflows: Otherwise, we have: From a risk management point of view, it follows that redemption shocks based on net redemptions may be underestimated compared to redemption shocks based on gross redemptions. However, we will see later that the approximation R (t) ≈ R − (t) may be empirically valid under some conditions. The computation of redemption rates only makes sense if they are homogeneous, coherent and comparable. Let us assume that we compute the redemption rate R (t) at the level of the asset management company, and we have the historical data for the last ten years. By assuming that there are 260 market days per year, we have a sample of 2 600 redemption rates. We can compute the mean, the standard deviation, different quantiles, etc. Does it help with building a stress scenario for a mutual fund? Certainly not, because redemptions depend on the specific investor behavior at the fund level and not on the overall investor behavior at the asset manager level. For instance, we can assume that an investor does not have the same behavior if he is invested in an equity fund or a money market fund. We can also assume that the redemption behavior is not the same for a central bank, a retail investor, or a pension fund. Therefore, we must build categories that correspond to homogenous behaviors. Otherwise, we will obtain categories, whose behavior is nonstationary. But, without the stationarity property, risk measurement is impossible and stress testing is a hazardous exercise. Therefore, liability categorization is an important step before computing redemption rates. For instance, ESMA (2019) considers four factors regarding investor behavior: investor category, investor concentration, investor location and investor strategy. Even though the last three factors are significant, the most important factor remains the investor type. For instance, AMF (2017, page 12) gives an example with the following investor types: large institutional (tier one), small institutional (tier two), investment (or mutual) fund, private banking network and retail investor. Other categories can be added: central bank, sovereign, corporate, third-party distributor, employee savings plan, wealth management, etc. Moreover, it is also important to classify funds into homogeneous buckets such as balanced funds, bond funds, equity funds, etc. An example of an investor/fund categorization matrix is given in Table 1 . The granularity of the investor/fund classification is an important issue. It is important to have a very detailed classification at the level of the database in order to group categories together from a computational point of view. In order to calibrate stress scenarios, we must have a sufficient number of observations in each cell of the classification matrix. Let us for instance consider the case of central banks. We can suppose that their behavior is very different to the other investors. Therefore, it is important for an asset manager to be aware of the liabilities with respect to central banks. Nevertheless, there are few central banks in the world, meaning we may not have enough observations for calibrating some cells (e.g. central bank/equity or central bank/real asset), and we have to merge some cells (across investor and fund categories). We consider a fund. We note TNA i (t) the assets under management of the investor i for this fund. By definition, we have: where NAV (t) is the net asset value of the fund and N i (t) is the number of units held by the investor i for the fund. The fund's assets under management are equal to: where TNA (k) (t) = i∈IC (k) TNA i (t), and IC (k) is the k th investor category. It follows that: is the total number of units in issue. We retrieve the definition of the assets under management (or total net assets) at the fund level. We can obtain a similar breakdown for the outflows 4 : The redemption rate for the investor category IC (k) is then equal to: We deduce that the relationship between the investor-based redemption rates and the fundbased redemption rate is: where ω (k) (t) represents the weights of the investor category IC (k) in the fund: Equation (9) is very important, because it shows that the redemption rate at the fund level is a weighted-average of the redemption rates of the different investor categories. Let us now consider different funds. We note R (f,k) (t) as the redemption rate of the investor category IC (k) for the fund f at time t. By relating the fund f to its fund category FC (j) , we obtain a database of redemption rates by investor category IC (k) and fund category FC (j) : is then the sample of all redemption rates of the investor category IC (k) for all the funds that fall into the fund category FC (j) during the observation period T . We notice that DB (j,k) (t) does not have a unique element for a given date t because we generally observe several redemptions at the same date for different funds and the same investor category. The key parameter for computing the redemption flows is the redemption rate, which is defined for an investor category and a fund category. It is not calibrated at the fund level, because past redemption data for a given specific fund are generally not enough to obtain a robust estimation. This is why we have pooled redemption data as described in the previous paragraph. Using these data, we can estimate the probability distribution F of the redemption rate and define several statistics that can help to build stress scenarios. In what follows, we consider the liability data provided by Amundi Asset Management from January, 1 st 2019 to August, 19 th 2020. The database is called 'Amundi Cube Database' and contains 1 617 403 observations if we filter based on funds with assets under management greater than e5 mn. The breakdown by investor categories 5 is given in In what follows, we apply a filter that consists in removing observations that corresponds to dedicated mutual funds (FCP and SICAV) and mandates (see Table 41 on page 93). The motivation is to focus on mutual funds with several investors, and this issue will be extensively discussed in Section 4.1.3 on page 44. We first begin by comparing the gross redemption rate R , the net flow rate R ± and the net redemption rate R − . Some results are given in Figures 4 and 5 for retail and insurance investors and bond and equity funds. In the case of insurance companies, we notice that the approximation R ≈ R − ≈ −R ± is valid when the redemption rate is greater than 20%. This is not the case for retail investors, where we observe that some large redemptions may be offset by large subscriptions 6 . The difference between retail and insurance categories lies in the investor concentration. When an investor category is concentrated, there is a low probability that this offsetting effect will be observed. This is not the case when the granularity of the investor category is high. We also observe that the approximation R ≈ R − ≈ −R ± depends on the fund category. For instance, it is not valid for money market funds. The reason is that we generally observe subscriptions in a bull market and redemptions in a bear market when the investment decision mainly depends on the performance of the asset class. This is why large redemptions and subscriptions tend to be mutually exclusive (in the mathematical sense) in equity or bond funds. The mutual exclusivity property is more difficult to observe for money market, enhanced treasury or balanced funds, because their inflows are less dependent on market conditions. We conclude that net redemption rates may be used in order to perform stress scenarios under some conditions regarding the concentration of the investor category and the type of mutual fund. For a given investor/fund category, we note F as the probability distribution of the redemption rate. We can define several risk measures (Roncalli, 2020 , pages 62-63): • the mean: • the standard deviation-based risk measure: • the quantile (or the value-at-risk) at the confidence level α: • the average beyond the quantile (or the conditional value-at-risk): The choice of a risk measure depends on its use. For instance, M can be used by the fund manager daily, because it is the expected value of the daily redemption rate. If the fund manager prefers to have a more conservative measure, he can use SD (1). M and SD (c) make sense in normal periods from a portfolio management perspective, but they are less relevant in a stress period. This is why it is better to use Q (α) and C (α) from a risk management point of view. In the asset management industry, the consensus is to set α = 99%. In Table 2 , we have reported the values of M, Q (99%) and C (99%) by considering the empirical distribution of gross redemption rates by client category. We do not consider the SD-measure because we will see later that there is an issue when it is directly computed from a sample of historical redemption rates. On average, the expected redemption rate is roughly equal to 20 bps. It differs from one client category to another, since the lowest value of M is observed for central banks whereas the highest value of M is observed for corporates. The 99% value-at-risk is equal to 3.5%. This means that we observe a redemption rate of 3.5% every 100 days, that is every five months. Again, there are some big differences between the client categories. The riskiest category is corporate followed by sovereign and auto-consumption. If we focus on the conditional value-at-risk, we are surprised by the high values taken by C (99%). If we consider all investor categories, C (99%) is more than 15%, and the ratio R (99%) between C (99%) and Q (99%) is equal to 4.51 . This is a very high figure since this ratio is generally less than 2 for market and credit risks. For example, in the case of a Gaussian distribution N 0, σ 2 , the ratio R (α) between the conditional value-at-risk and the value-at-risk is equal to: This ratio is respectively equal to 1.37 and 1.15 when α = 90% and α = 99%. Moreover, Roncalli (2020, page 118) showed that it is a decreasing function of α and: We deduce that the ratio is lower than 1.5 for reasonable values of the confidence level α. Therefore, the previous figure R (99%) = 4.51 indicates that redemption risk is more skewed than market and credit risks. Table 3 reports the statistical measures by fund category. Again, we observe some big differences. Money market and enhanced treasury funds face a high redemption risk followed by bond and equity funds. This is normal because treasury funds can be converted to cash very quickly, investors are motivated to redeem these funds when they need cash, and their holding period is short. At the global level, we also notice that the redemption behavior is similar between bond and equity funds. For instance, their 99% value-at-risk is close to 3% (compared to 6% for the enhanced treasury category and 22% for money market funds). Another interesting result is the lower redemption rate of balanced funds compared to bond and equity funds. This result is normal because balanced funds are more diversified. Therefore, investors in balanced funds are more or less protected by a bond crisis or an equity crisis. Finally, structured funds are the least exposed category to redemption risk, because they generally include a capital guarantee or protection option. 32 3.52 15.79 (1) = balanced, (2) = bond, (3) = enhanced treasury, (4) = equity, (5) = money market, (6) = other, (7) = structured, (8) = total The historical statistical measures 7 for the classification matrix are given in Tables 4, 5 and 6. We notice that the two dimensions are important, since one dimension does not dominate the other. This means that a low-risk (resp. high-risk) investor category tends to present the lowest (resp. highest) redemption statistics whatever the fund category. In addition, the ranking of redemption statistics between fund categories is similar whatever the investor category. Nevertheless, we observe some exceptions and new stylized facts. For instance, we have previously noticed that bond and equity funds have similar redemption rates on average. This is not the case for the corporate, corporate pension fund and sovereign categories, for which historical C-statistics are more important for equity funds than bond funds. For the corporate pension fund category, the risk is also higher for balanced funds than for bond funds. According to BCBS (2017, page 60), a historical stress scenario "aims at replicating the changes in risk factor shocks that took place in an actual past episode". If we apply this definition to the redemption risk, the computation of the historical stress scenario is simple. First, we have to choose a stress period T stress and second, we compute the maximum redemption rate: X T stress = max t∈T stress R (t) For example, if we apply this definition to our study period, we obtain the results given in Table 7 . We recall that the study period runs from January 2019 to August 2020 and includes the Coronavirus pandemic crisis, which was a redemption stress period. We observe that the X-statistic is generally equal to 100%! This is a big issue, because it is not helpful to consider that liquidity stress testing of liabilities leads to a figure of 100%. The problem is that the X-statistic is not adapted to redemption risk. Let us consider an investor category IC (k) and a fund category FC (j) . The X-statistic is computed by taking the maximum of all redemption rates for all funds that belong to the fund category: If there is one fund with only one investor and if this investor redeems 100%, X (j,k) (T stress ) is equal to 100%. However, the asset manager does not really face a liquidity risk in this situation, because there is no other investor in this fund. So, the other investors are not penalized. We have excluded this type of fund. However, we face a similar situation in many other cases: small funds with a large fund holder, funds with a low number of unitholders, etc. Moreover, this type of approach penalizes big asset managers, which have hundreds of funds. Let us consider an example. For a given investor/fund category, the fund manager A has 100 funds of $100 million, whereas the fund manager B has one fund of $10 billion. From a theoretical point of view, A and B face the same redemption risk, since they both manage $10 billions for the same investor/fund category. However, it is obvious that X A X B , meaning that the historical stress scenario for the fund manager A will be much higher than the historical stress scenario for the fund manager B. This is just a probabilistic counting principle as shown in Appendix A.1 on page 78. If we consider the previous example, the historical stress scenario for the fund manager A is larger than 99.9% when the historical stress scenario for the fund manager B is larger than 6.68% (see Figure 38 on page 95). More generally, the two stress scenarios are related in the following manner: where X 1 is the X-measure for one fund and X n is the X-measure for n funds. (1) = balanced, (2) = bond, (3) = enhanced treasury, (4) = equity, (5) = money market, (6) = other, (7) = structured Remark 4 Another approach consists in computing the average redemption rate daily: where the weights are proportional to the size of funds f that belong to the j th fund category FC (j) . In this case, we have: This method does not have the previous drawback, but it has other shortcomings such as an information loss. However, the biggest disadvantage is that the historical stress scenario is generally based on the largest fund, except when the funds have similar size. Since X-measures can not be used to build redemption shocks, we propose using Q or C-measures. Q (99%) is the daily value-at-risk at the 99% confidence level. This means that its return period is 100 days. On average, we must observe that redemption shocks are greater than Q (99%) two and a half times per year. We can also use the conditional valueat-risk C (99%) if we want more severe redemption shocks. The drawback of C (99%) is that we don't know the return period of such event. However, it does make sense because it is a very popular measure in risk management, and it is well received by regulatory bodies and supervisors (Roncalli, 2020) . Nevertheless, we must be cautious about the computed figures obtained in Tables 5 and 6 on page 18. For example, we don't have the same confidence level between the matrix cells, because the estimates are not based on the same number of observations. In the case of retail investors or third-party distributors, we generally use a huge number of observations whereas this is not the case with the other categories. In Table 8 , we give an example of confidence level codification. We see that some cells are not well estimated since the number of observations is less than 10 000. For some of them, the number of observations is very low (less than 200), implying that the confidence on these estimates is very poor. Therefore, the estimated values cannot be directly used as redemption shocks. However, they help risk managers and business experts to build redemption shocks. Starting from these figures, they can modify them and build a table of redemption shocks that respect the risk coherency C investor between investor categories and the risk coherency C fund between fund categories 8 . The risk coherency C investor means that if one investor category is assumed to be riskier than another, the global redemption shock of the first category must be greater than the global redemption shock of the second category: For example, if we consider the Q-measure, we can propose the following risk ordering: 5. auto-consumption, corporate, sovereign In this case, the redemption shock S (j,k) for the (j, k)-cell depends on the global redemption shock S (k) for the investor category IC (k) . For instance, we can set the following rule of thumb: where m (j) is the multiplicative factor of the fund category FC (j) . In a similar way, the risk coherency C fund means that if one fund category is assumed to be riskier than another, the global redemption shock of the first category must be greater than the global redemption shock of the second category: For example, if we consider the Q-measure, we can propose the following risk ordering: The redemption shock S (j,k) for the (j, k)-cell depends then on the redemption shock S (j) for the fund category IC (j) . Again, we can set the following rule of thumb: where m (k) is the multiplicative factor of the investor category IC (k) . We can also combine the two rules of thumb and we obtain the mixed rule: Let us illustrate the previous rules of thumb by considering the Q-measure. Table 9 gives an example of S (j,k) by considering the risk coherency 9 C investor , whereas Table 10 corresponds to the risk coherency 10 C fund . The mixed rule is reported in Table 11 . These figures can then be modified by risk managers and business experts by considering the specificity of some matrix cells. For instance, it is perhaps not realistic to have the same redemption shock for balanced funds between auto-consumption and corporates. Moreover, these redemption shocks can also be modified by taking into account the C-measure. For instance, the conditional value-at-risk for bond funds is much higher for third-party distributors than for sovereigns. Perhaps we can modify the redemption shock of 3.3% and have a larger value for third-party distributors. It is even more likely that the estimated values of Q and C are based on 75 591 observations for the third-party distributor category, and 2 261 for the sovereign category. Therefore, we can consider that the estimated value of 4.59% obtained in Table 5 on page 18 does make more sense than the proposed value of 3.3% obtained in Table 11 for the third-party distributor/bond matrix cell. In a similar way, we can consider that the estimated value of 0.14% does make less sense than the proposed value of 7.0% for the sovereign/bond matrix cell. The previous analysis shows that building redemption shocks in a stress testing framework is more of an art than a science. A pure quantitative approach is dangerous because it is data-driven and it does not respect some coherency properties. However, historical statistics are very important because they provide an anchor point for risk managers and business experts in order to propose stress scenarios that are satisfactory from regulatory, risk management and fund management points of view. Historical data are also important because they help to understand the behavior of clients. It is different from one fund category to another, it also depends on the granularity of the classification, it may depend on the time period, etc. In what follows, we complete this pure historical analysis using more theoretical models. These models are important, because an historical approach is limited when we want to understand contagion effects between investors, correlation patterns between funds, time properties of redemption risk, the impact of the holding period, etc. The idea is not to substitute one model with another, but to rely on several approaches, because there is not just one single solution to the liability stress testing problem. 9 We use the following values: S (k) = 0.5% for central banks and corporate pension funds, S (k) = 2% for employee savings plans and retail, S (k) = 3.5% for other and third-party distributors, S (k) = 5% for institutionals and insurance companies, and S (k) = 8% for auto-consumption, corporates and sovereigns. For the multiplication factor, we assume that m (j) = 0.25 for structured, m (j) = 0.5 for balanced and other, m (j) = 1 for bond and equity, m (j) = 1.75 for enhanced treasury, and m (j) = 6 for money market. 10 We use the following values: S (j) = 0.5% for structured, S (j) = 1.5% for balanced and other, S (j) = 3% for bond and equity, S (j) = 5% for enhanced treasury, and S (j) = 20% for money market. For the multiplication factor, we assume that m (k) = 0.25 for central banks and corporate pension funds, m (k) = 0.5 for employee savings plans and retail, m (k) = 1 for other and third-party distributors, m (k) = 1.5 for institutionals and insurance companies, and m (k) = 2 for auto-consumption, corporates and sovereigns. (10) (1) (2) (3) (4) (5) (6) (7) (11) (1) (2) (3) (4) (5) (6) (7) (8) Table 11 : Redemption shocks in % computed with the rule of thumb (12) (1) (2) (3) (4) (5) (6) (7) (8) Auto-consumption 3.5 (1) = balanced, (2) = bond, (3) = enhanced treasury, (4) = equity, (5) = money market, (6) = other, (7) = structured, (8) = total 3 The frequency-severity modeling approach The direct computation of value-at-risk, conditional value-at-risk and other statistics from historical redemption rates is particularly problematic. Indeed, we observe a large proportion of zeros in the redemption rate database. On average, we have 68.9% of zeros, this proportion reaches 99.5% for some investors and it is more than 99.9% for some matrix cells. Therefore, the data of redemption rates are "clumped-at-zero", meaning that the redemption rate is a semi-continuous random variable, and not a continuous random variable (Min and Agresti, 2002) . This discontinuity is a real problem when estimating the probability distribution F. This is why we consider that the redemption rate is not the right redemption risk factor. We prefer to assume that the redemption risk is driven by two dimensions or two risk factors: 1. the redemption frequency, which measures the occurrence E of the redemption; 2. the redemption severity R , which measures the amount of the redemption. It is obvious that this modeling approach finds its root in other risk models that deal with extreme events or counting processes, such as operational and insurance risks (Roncalli, 2020). In the frequency-severity approach, we distinguish the redemption event E that indicates if there is a redemption, and the redemption amount R that measures the redemption rate in case of a redemption. An example is provided in Figure 7 . The probability to observe a redemption is equal to 5%, and in the case of a redemption, the amount can be 2%, 5%, 15% and 50%. It follows that the redemption rate is the convolution of two risk factors. We assume that the redemption event E follows a bernoulli distribution B (p), whereas the redemption severity 11 R follows a continuous probability distribution G. We have: We deduce that the unconditional probability distribution of the redemption rate is given by: Its density probability function is singular at x = 0: is the density function of G. Some examples are provided in Figure 8 when p = 5%. We observe that the density function is composed of a dirac measure and a continuous function. In the case of G 1 , the distribution is right-skewed, meaning that the probability to observe small redemptions is high. In the case of G 2 , we have a bell curve, meaning that the redemption amount is located around the mean if there is a redemption. Finally, the distribution is left-skewed in the case of G 3 , meaning that the probability to observe high redemptions is high if there is of course a redemption, because we recall that the probability to observe a redemption is only equal to 5%. From a probabilistic point of view, the redemption rate is then the product of the redemption event and the redemption severity: In Appendix A.2.1 on page 78, we show that: and: Moreover, the skewness coefficient is equal to: where: For the excess kurtosis coefficient, we obtain: where: In Figure 9 we have reported the moments of the redemption rate R by considering the following set of parameters: We notice that the parameter values of R have a major impact on the statistical moments, but the biggest effect comes from the frequency probability p. Indeed, we verify the following properties: This means that the redemption risk is very high for small frequency properties. In this case, the expected redemption rate and its standard deviation are very low, but skewness and kurtosis risk are very high! This creates a myopic situation where the asset manager may have the feeling that redemption risk is not a concern because of historical data. Indeed, when p is low, the probability of observing large redemption rates is small, implying that they are generally not observed in the database. For instance, let us consider two categories that have the same redemption severity distribution, but differ from their redemption frequency probability. One has a probability of 50%, the other has a probability of 1%. It is not obvious that the second category experienced sufficient severe redemption events such that the historical data are representative of the severity risk. For the M-measure, we have: The formula of the value-at-risk is equal to: We notice that computing the quantile α of the unconditional distribution F is equivalent to compute the quantile α G of the severity distribution G: The relationship between p, α and α G is illustrated in Figure 39 on page 95. Let us focus on the 99% value-at-risk: If the redemption frequency probability is greater than 1%, the value-at-risk corresponds to the quantile (p − 1%) /p. The relationship between p and α G = (p − 1%) /p is shown in Figure 10 . If p is greater than 20%, α G is greater than 95%. If p is less than 5%, we observe a high curvature of the relationship, implying that we face a high estimation risk. For instance, if p is equal to 1.5%, the 99% value-at-risk corresponds to the quantile 3.33% of the redemption severity. If p becomes 2.0%, the 99% value-at-risk is then equal to the quantile 50% of the redemption severity! Therefore, there is a high sensitivity of the 99% value-at-risk when p is low, implying that a small error in the estimated value of p leads to a high impact on the value-at-risk. For the conditional value-at-risk, we obtain: where Q (u) is the quantile function of R for the confidence level u. In the case where p > 1 − α, we obtain: Another expression of the conditional value-at-risk is: In the case where p > 1 − α, we obtain: where g (x) is the probability density function of G (x). All these formulas can be computed numerically thanks to Gauss-Legendre integration. We now introduce a new risk measure which is very popular when considering parametric model. Roncalli (2020) defines the distribution-based (or parametric-based) stress scenario Liquidity Stress Testing in Asset Management S (T ) for a given horizon time T such that the return time of this scenario is exactly equal to T . From a mathematical point of view, we have: Pr {R ≥ S (T )} is the exceedance probability of the stress scenario, implying that the quantity Pr {R ≥ S (T )} −1 is the return time of the exceedance event. For example, if we set The return time associated to a 99% value-at-risk is then equal to 100 days, the return time associated to a 99.9% value-at-risk is equal to 1 000 days (or approximately 4 years), etc. This parametric approach of stress testing is popular among professionals, regulators and academics when they use the extreme value theory for modeling the risk factors. By combining the two definitions S (T ) = Q (α) and T = (1 − α) −1 , we obtain the mathematical expression of the parametric stress scenario: If we consider the zero-inflated model, we deduce that: The magnitude of T is the year, but the unit of T is the day. For example, since one year corresponds to 260 market days, the five-year stress scenario is equal to 12 : The choice of the severity distribution is an important issue. Since R is a random variable between 0 and 1, it is natural to use the two-parameter beta distribution B (a, b). We have: is the incomplete beta function. The corresponding probability density function is equal to: where B (a, b) is the beta function: Concerning the statistical moments, the formulas are given in Appendix A.2.2 on page 81. We report some examples of density function in Figure 11 . Instead of providing the parameters a and b, we have indicated the value µ and σ of the mean and the volatility. The first distribution is skewed, because the volatility is high compared to the mean. The other three distributions have a mode. Figure 12 shows the corresponding statistical moments of Liquidity Stress Testing in Asset Management the associated zero-inflated model. We notice that the first and third distributions have the largest skewness and kurtosis. In Figure 13 , we report the 99% value-at-risk of the redemption rate. As explained before, the Q-measure highly depends on the redemption frequency p. Again, we observe that the sensitivity of the value-at-risk is particularly important when p is small 13 . The ratio between the 99% conditional value-at-risk and the 99% value-at-risk is given in Figure 14 . When the redemption frequency p is high, the ratio is less than 1.5 and we retrieve the typical figures that we observe for market and credit risks 14 . When the redemption frequency p is small, the ratio may be greater than 2.0. These results shows that the sensitivity to redemption risk is very high when the observed redemption frequency is low. The stress scenarios S (T ) are given in Figure 15 when the redemption frequency p is equal to 1%. By definition, S (T ) increases with the return time T . From a theoretical point of view, the limit of the stress scenario is 100%: However, we observe that stress scenarios reach a plateau at five years, meaning that stress scenarios beyond 5 years have no interest. This is true for small values of p, but it is even more the case for larger values of p as shown in Figures 40 and 41 on page 96. Remark 5 In order to better understand the use of the C-measure as a stress scenario, we compute the implied return time such that the stress scenario is exactly equal to the 12 We assume that the redemption frequency is greater than 1/1300 or 7.69 bps. Otherwise, the quantile is equal to zero. 13 Because of the impact of p on the confidence level α G -see Figure 10 on page 28. 14 When p tends to one, the ratio is respectively equal to 1.15, 1.09, 1.06 and 1.03 for the four probability distributions of the redemption severity. conditional value-at-risk: Results are given in Table 12 . We notice that the value is between 0.77 and 1. 03 . On average, we can consider that the return time of the 99% conditional value-at-risk is about one year. This is 2.6 times the return time of the 99% value-at-risk 15 . The choice of the beta distribution is natural since the support is [0, 1], but we can consider other continuous probability distributions for modeling R . For example, the Kumaraswamy distribution is another good candidate, but it is close to the beta distribution. When the support of the probability distribution is [0, ∞), we apply the truncation formula 16 : For instance, we can use the gamma or log-logistic distribution. However, our experience shows that some continuous probability distributions are not adapted such as the log-gamma and log-normal distributions, because the logarithm transform performs a bad scale for random variables in [0, 1]. Finally, we can also use the logit transformation, which is very popular for modeling the probability of default (PD) or the loss given default (LGD) in credit risk. Following Roncalli (2020, page 910), we assume that R is a logit transformation of a random variable X ∈ (−∞, ∞), meaning that 17 : For instance, in the case of the logit-normal distribution, we have: We recall that the return time of the 99% value-at-risk is equal to 100 market days or 100 260 ≈ 0.38 years. 16 For the probability density function, we have: 17 We also have: and: A summary of these alternative approaches 18 is given in Table 13 . In the sequel, we continue to use the beta distribution, because it is easy to calibrate and it is the most popular approach when modeling a random variable in [0, 1]. However, we cannot claim that it is the best fitting model. Such a debate has already taken place in operational risk with the lognormal distribution and the modeling of the severity distribution of operational risk losses (Roncalli, 2020) . Nevertheless, we think that this debate is too early in the case of liability stress testing, and can wait when we will have more comprehensive redemption databases. As explained previously, the zero-inflated beta model is appealing for producing stress scenarios. For that, we proceed in two steps. We first calibrate the parameters of the model, and then we compute the stress scenarios for a given return time. Let Ω = {R 1 , . . . , R n } be the sample of redemption rates for a given matrix cell. Three parameters have to be estimated: the redemption frequency p and the parameters a and b that control the shape of the beta distribution. We note n 0 as the number of observations that are equal to zero and n 1 = n − n 0 as the number of observations that are strictly positive 19 . In Appendix A.3 on page 81, we show that the maximum likelihood estimates are:p = n 1 n 0 + n 1 18 γ (α, x) is the lower incomplete gamma function. 19 We have and: The estimatesâ andb can be found by numerical optimization. This is the traditional approach for estimating a zero-inflated model. However, it is not convenient since the parameters p,â,b should be modified by risk managers and business experts before computing redemption shocks. Indeed, the calibration process of parametric stress scenarios follows the same process when one builds historical stress scenarios, and estimated values p,â,b cannot be directly used because they do not necessarily respect some risk coherency principles and their robustness varies across matrix cells. A second approach consists in using the method of moments. In this case, the estimator of p has the same expression:p = n 1 n 0 + n 1 For the parameters of the beta distribution, we first calculate the empirical meanμ and the standard deviationσ of the positive redemption rates R , and then we use the following relationships (Roncalli, 2020, page 193) : The differences between the two methods are the following: • In the case of the method of maximum likelihood, a and b are explicit parameters. Once the parameters p, a and b are estimated, we can calculate the mean µ and standard deviation σ for the severity distribution. In this approach, µ and σ are implicit, because they are deduced from a and b. • In the case of the method of moments, a and b are implicit parameters. Indeed, they are calculated after having estimated the mean µ and standard deviation σ for the severity distribution. In this approach, µ and σ are explicit and define the severity distribution. The first approach is known as the p − a − b parameterization, whereas the second approach corresponds to the p − µ − σ parameterization. By construction, this last approach is more convenient in a liquidity stress testing framework, because the parameters µ and σ are intuitive and self-explanatory measures, which is not the case of a and b. Therefore, they can be manipulated by risk managers and business experts. We have estimated the parameters p, a, b, µ and σ with the two methods. Table 14 shows the redemption frequency. On average,p is equal to 31%, but we observe large differences between the matrix cells. For instance,p is less than 5% for central banks, corporate pension funds and employee savings plans, whereas the largest values ofp are observed for retail investors and third-party distributors. The values ofμ andσ are reported in Tables 15 and 16. The average redemption severity is 0.72%, whereas the redemption volatility is 4.55%. Again, we observe some large differences between the matrix cells. Tables 42 and 43 (1) = balanced, (2) = bond, (3) = enhanced treasury, (4) = equity, (5) = money market, (6) = other, (7) = structured, (8) = total Using the previous estimates (p,μ,σ), risk managers and business experts can define the triplet (p, µ, σ) for the different matrix cells. For that, they must assess the confidence in estimated values with respect to the number of observations. For the frequency parameter, we use the value of n, which has been already reported in Table 8 on page 21 . For the severity parametersμ andσ, we use the value of n 1 , which is much smaller than n. Using the data given in Table 41 on page 93, we have built the confidence measure in Table 17 . We confirm that the confidence measure inμ andσ is lower than the confidence measure inp. In particular, there are many matrix cells, where the number n 1 of observations is lower than 200. This explains why Tables 15 and 16 contain a lot of missing values. Therefore, except for a few matrix cells, the estimated valuesμ andσ must be challenged by risk managers and business experts. Again, they can use risk coherency principles 20 C investor and C fund to build their own figures of p, µ and σ. Once the triplet (p, µ, σ) is defined for each matrix cell, we compute stress scenarios using the following formula: where B −1 (α; a, b) is the α-quantile of the beta distribution with parameters a and b. The parametric stress scenario S (T ; p, µ, σ) depends on the return time T and the three parameters of the zero-inflated model. An example is provided in Figure 16 . For each plot, we indicate the triplet (p, µ, σ). For instance, the first plot corresponds to the triplet (2%, 1%, 2%), meaning that the daily redemption frequency is 2%, the expected redemption severity is 1% and the redemption volatility is 2%. In particular, these plots illustrate the high impact of σ, which is the key parameter when computing parametric stress scenarios. The reason is that the parameters p and µ determine the mean E [R ], whereas the uncertainty around this number is mainly driven by the parameter σ. The redemption volatility controls then the shape of the probability distribution of the redemption rate (both the skewness and the kurtosis), implying that σ has a major impact on the stress scenario S (T ) when T is large. In this section, we go beyond the zero-inflated model by considering the behavior of each investor. In particular, we show that the redemption rate depends on the liability structure of the mutual fund. Moreover, the behavior of investors may be correlated, in particular in a stress period. In this case, the modeling of spillover effects is important to define stress scenarios. The individual-based model and the zero-inflated model are highly connected. Indeed, the zero-inflated model can be seen as a special case of the individual-based model when we summarize the behavior of all unitholders by the behavior of a single client. Let TNA (t) be the assets under management of an investment fund composed of n clients: where TNA i (t) is the net asset of the individual i. The redemption rate of the fund is equal to the redemption flows divided by the total net assets: where ω i represents the weight of the client i in the fund: Generally, we assume that the clients are homogenous, meaning that E i and R i are iid random variables. If we denote byp andG the individual redemption probability and the individual severity distribution. The individual-based model is then characterized by the 4-tuple n, ω,p,G , where n is the number of clients and ω = (ω 1 , . . . , ω n ) is the vector of weights. Like the zero-inflated model, we consider aμ −σ parameterization ofG, meaning that the model is denoted by IM (n, ω,p,μ,σ). Remark 7 When the individual severity distributionG is no specified, we assume that it is a beta distribution B ã,b , whose parametersã andb are calibrated with respect to the severity meanμ and the severity volatilityσ using the method of moments. In a similar way, we assume that the vector of weights is equally-weighted when it is not specified. In this case, the individual-based model is denoted by IM (n,p,μ,σ). In Figure 17 , we report the histogram of the redemption rate R in the casep = 50%, µ = 50% andσ = 10%. In the case n = 1, we obtain a singular distribution. Indeed, there is a probability of 50% that there is no redemption. The singularity decreases with respect to the number n of investors, because the probability to have a redemption increases. This is normal since the redemption frequency of a mutual fund depends on the number of unitholders. This explains that the redemption frequency is larger for a retail fund than for an institutional fund. The skewness effect The singularity of the distribution function F at the point R = 0 is entirely explained by the two parametersp and n as shown in Appendix A. 4 .1 on page 83, because we have: The fact that the probability distribution is not continuous has an impact on the skewness and the kurtosis. In Table 18 , we have reported the probability Pr {R = 0}. If there is a few investors in the fund, the probability to observe no redemption in the fund is high. For instance, ifp = 5% and n = 10, we have Pr {R = 0} = 59.87%. Ifp = 1% and n = 10, we have Pr {R = 0} = 90.44%. How to interpret these results? Sincep is the individual redemption probability, 1/p is the return time of a redemption at the investor level. For example,p = 5% (resp.p = 1%) means that investors redeem every 20 days (resp. 100 days) on average. At the fund level, the return time to observe a redemption is equal to (1 − Pr {R = 0}) −1 . For instance, in the casep = 5% and n = 10, we observe a redemption two days per week in the fund on average 21 . This analysis may help to distinguish active and passive investors. In the case of passive investors when the redemption event occurs once a year or less,p is less than 40 bps. In the case of active investors that redeem once a month,p is greater than 5%. Therefore, the skewness effect depends if the fund has active investors or not, and if the fund is granular or not. The mean effect The mean shape is easy to understand since it is the product of the redemption probability and the mean of the redemption severity: Curiously, it depends neither on the number of investors in the fund, nor on the liability structure (see Figure 18 ). Sinceμ ∈ [0, 1], we notice that E [R ] ≤p, meaning that we must observe very low values of the redemption mean. And we verify this property if we consider the results 22 given in Table 4 on page 18. If we consider all investor and fund categories, the mean is equal to 22 bps. The largest value is observed for the sovereign/money market category and is equal to 1.91%. The volatility effect By assuming that the liability weights are equal (ω i = 1/n), the volatility of the redemption rate is equal to: Globally, we observe that σ 2 (R ) is an increasing function ofp,μ andσ as shown in Figure 19 . When the redemption probability increases, we observe a convexity shape because we have: n However, this is not realistic sincep ≤ 20% in practice. Another interesting property is that σ 2 (R ) tends to zero when the number of investors in the fund increases ( Figure 20) . If we compute the volatility of the redemption rate, we obtain the figures given in Table 48 on page 100. We observe that σ (R ) E [R ], implying that R is a high-skewed random variable. This challenges the use of the SD (c) measure presented on page 16. Correspondence between zero-inflated and individual-based models We notice that the zero-inflated model ZI (p, µ, σ) is a special case of the individual-based model by 22 Another way to compute the empirical mean of R is to calculate the product of the aggregate redemption frequency p (Table 14 on page 36) and the aggregate severity mean (Table 15 on considering only one unitholder. Therefore, it is obvious that the zero-inflated model can not be seen as an explanatory model. It is a reduced-form model or a parametric model that can fit the data, but the interpretation of the p − µ − σ parameterization is not obvious, because ZI (p, µ, σ) is an aggregate population model. In this paragraph, we would like to find the relationships between the parameters of the zero-inflated model and those of the individual-based model, such that the two models are statistically equivalent: ZI (p, µ, σ) ≡ IM (n, ω,p,μ,σ) There are different approaches. A first one is to minimize the Kolmogorov-Smirnov statistics between ZI (p, µ, σ) and IM (n, ω,p,μ,σ). Another approach consists in matching their moments. We consider the second approach because we obtain analytical formulas, whereas the solution of the first approach can only be numerical. In Appendix A.5 on page 85, we show that: whereas the parameter σ satisfies the following relationship: where H (ω) = n i=1 ω 2 i is the Herfindahl index. It is interesting to notice that p is a function of n andp, µ is a function of n,p andμ, but σ does not only depends on the parameters n, p,μ andσ:    p = ϕ 1 (n,p) µ = ϕ 2 (n,p,μ) σ = ϕ 3 (n,p,μ,σ, H (ω)) Indeed, the aggregate severity volatility also depends on the Herfindahl index of the fund liability structure. The previous relationships can be inverted in order to define the parameters of the individual-based model with respect to the parameters of the zero-inflated model: However, we notice that there are two degrees of freedom -n and H (ω) -that must be fixed. In Tables 19 and 20 , we report some examples of calibration when n is equal to 10 and ω i is equal to 10%. For instance, if the parameters of the individual-based model arẽ p = 1.00%,μ = 50% andσ = 10%, we obtain p = 9.56%, µ = 5.23% and σ = 1.48% for the zero-inflated model. If we know the weights of the investors in the investment fund, we can therefore calibrate the zero-inflated model from the individual-based model (Table 19) , but also the individual-based model from the zero-inflated model (Table 20) . We notice that the variance of the redemption rate depends on the Herfindahl index: This implies that the liability structure ω is an important parameter to understand the probability distribution of the redemption rate. The arithmetics of the Herfindahl index We know that the Herfindahl index is bounded: 1 n ≤ H (ω) ≤ 1 H (ω) is equal to one when one investor holds 100% of the investment fund (∃i : ω i = 1), whereas the lower bound is reached for an equally-weighted liability structure (ω i = n −1 ). Therefore, H (ω) is a measure of concentration. A related statistic is the "effective number of unitholders": N (ω) indicates how many equivalent investors hold the investment fund. For instance, we consider two funds with the following liability structures ω (1) = (25%, 25%, 25%, 25%) and ω (2) = (42%, 17%, 15%, 13%, 9%, 3%, 1%). Since we have N ω (1) = 4 and N ω (2) = 3.94, we may consider that the first fund is not more concentrated than the second fund even if the second fund has 7 unitholders. We assume that the liability weights follow a geometric series with ω i ∝ q i and 0 < q < 1. We have 23 : (1 − q) 2 23 Because we have: As shown in Figure 21 , we have an infinite number of unitholders, but a finite number of effective unitholders. For example, if q ≤ 0.98, then N (ω) < 100. Approximation of the probability distributionF (x | ω) We recall that the unconditional probability distribution of the redemption rate is given by F (x) = Pr {R ≤ x}. Since the redemption rate depends on the liability structure ω in the individual-based model IM (n, ω,p,μ,σ), we noteF (x | ω) the associated probability distribution: We now consider the model IM (N (ω) ,p,μ,σ) and defineF (x | H) as follows: The issue is to know under which conditions we can approximateF (x | ω) byF (x | H). Let us consider some Monte Carlo experimentations. We assume that the liability weights are geometric distributed: ω i ∝ q i . In Figure 22 , we compare the two probability distribu-tionsF (x | ω) andF (x | H) for several sets of parameters 24 (p,μ,σ). The weights ω i for q = 0.95 are given in Figure 42 on page 100. We notice that the approximation ofF (x | ω) byF (x | H) is good and satisfies the Kolmogorov-Smirnov test at the 99% confidence level. This is not the case if we assume that q = 0.90 or q = 0.50 (see Figures 43 and 44 on page 101). To better understand these results, we assume thatp = 0.3,μ = 0.5 andσ = 0. 4 . When q is equal to 0.50, the effective number of unitholders is low and is equal to 3. In this case, the probability distributionF (x | H) is far from the probability distributionF (x | ω) as shown in Figure 23 . In fact, this case corresponds to an investment fund which is highly concentrated. The risk is then to observe redemptions from the largest unitholders. In particular, we notice thatF (x | H) presents some steps. The reason is that the redemption rate can be explained by the redemption of one unitholder, the redemption of two unitholders or the redemption of three unitholders. If we assume that q is equal to 0.90, the effective number of unitholders is larger and becomes 38. In this case, the probability distributioñ F (x | H) is close to the probability distributionF (x | ω), because the step effects disappear (see Figure 24 ). To summarize, the approximation ofF (x | ω) byF (x | H) cannot be good when the effective number of unitholders (or H (ω) −1 ) is low. Remark 9 In many cases, we don't know the comprehensive liability structure ω, but only the largest weights. In Appendix A.6 on page 86, we derive an upper bound H + m (ω) of the Herfindahl index H (ω) from the m largest weights. Therefore, we can deduce a lower bound of the effective number of unitholders: An example is provided in Table 21 when we assume that ω i ∝ q i . When the fund is highly concentrated, we obtain a good approximation of N (ω) with the 10 th or 20 th largest 1 unitholders. Otherwise, N (ω) is underestimated. However, this is not a real issue because we can think that generated stress scenarios will be overestimated. Indeed, using a lower value N (ω) increases σ (R ) as shown in Figure 20 on page 42, implying that the redemption risk is generally overestimated. Stress scenarios based on the largest unitholders The previous results show that the main risk in a concentrated fund comes from the behavior of the largest unitholders. It justifies the use of stress scenarios based on the order statistics ω k:n : min ω i = ω 1:n ≤ · · · ≤ ω k:n ≤ ω k+1:n ≤ · · · ≤ ω n:n = max ω i Then, we can define the stress scenario that corresponds to the full redemption of the m largest unitholders: S (m) = m k=1 ω n−k+1:n An example is given in Table 22 when the liability structure is defined by ω i ∝ q i . Of course, these stress scenarios S (m) make sense only if the fund presents some liability concentration. Otherwise, they are not informative. Using collective and mutual funds The calibration of the individual-based model is much more complicated than the calibration of the zero-inflated model. The reason is that it depends on the liability structure of the funds, which are not necessarily the same for the different funds. Let us consider the case of a single fund f . We can estimate the parameters p,μ andσ using the quadratic criterion: {p ,μ ,σ } = arg min p p (f ) − 1 + (1 −p) wherep (f ) ,μ (f ) andσ (f ) are the empirical estimates of the parameters p, µ and σ, and H (f ) is the Herfindahl index associated with the fund. In practice, the liability structure changes every day, meaning that the Herfindahl index is time-varying. Therefore, we can use the average of Herfindahl indices. The weights p , μ and σ indicate the relative importance of each moment condition. If we consider several funds, the previous criterion becomes: where (f ) is the relative weight of the fund f . In practice, the estimatesp ,μ andσ are very sensitive to the Herfindahl index because of the first and third moment conditions. To illustrate this point, we consider the institutional category and we assume that there is only one fund. On page 36, we found thatp (f ) = 8.23%,μ (f ) = 3.23% andσ (f ) = 10. 86% . If H (f ) = 5, we obtainp = 1.70%, µ = 15.61% andσ = 53. 31% . If H (f ) = 20, we obtainp = 0.43%,μ = 62.04% and σ = 212. 48% . In the case of the retail category, we found thatp (f ) = 45.61%,μ (f ) = 0.33% andσ (f ) = 2. 88% . If H (f ) = 1 000, we obtainp = 0.06%,μ = 247% andσ = 2 489%. If H (f ) = 10 000, we obtainp = 0.01%,μ = 2 472% andσ = 24 891%. These results are not realistic sinceμ > 1 andσ > 1. Using mandates and dedicated funds Collective investment and mutual funds are pooled investment vehicles, meaning that they are held by several investors. We now consider another type of funds with a single unitholder. They correspond to mandates and funds that are dedicated to a unique investor. In this case, the Herfindahl index is equal to one, and the solution of Problem (26) corresponds to the parameter set of the zero-inflated model: In our database, we can separate the observations between collective and mutual funds on one side and mandates and dedicated funds on the other side. In Tables 23, 24 and 25, we have estimated the parametersp,μ andσ by only considering mandates and dedicated funds. These results highly differ than those obtained for collective and mutual funds (Tables 14, 15 and 16 on page 36). First, we can calibrate a smaller number of cells. Indeed, we recall that the estimates are not calculated if the number of observations is less than 200. Second, the magnitude of the estimates is very different. If we consider all fund and investor categories, we obtainp = 3.34%,μ = 2.13% andσ = 10.27%, whereas we have previously found p = 31.11%, µ = 0.72% and σ = 4.55%. As expected, we verify thatp p andσ σ because of the following reasons: • the redemption probability is larger in a collective fund than in a dedicated fund because they are several investors; • the volatility of the redemption severity is smaller in a collective fund than in a dedicated fund because the behavior of the different investors is averaged, implying that the dispersion of redemption is reduced. (1) = balanced, (2) = bond, (3) = enhanced treasury, (4) = equity, (5) = money market, (6) = other, (7) = structured, (8) = total Curiously, we do not observe thatμ ≈ µ. One explanation may be that investors in mandates are not the same as investors in collective funds. Indeed, we may consider that they are more sophisticated and bigger when they are able to put in place a mandate or a dedicated fund. For instance, they can be more active. The results obtained with data from mandates and dedicated funds are more realistic than those obtained with data from collective and mutual funds. The drawback is that they are based on a smaller number of observations and there are many cells where we don't have enough observations for computing the estimates. Therefore, these estimates must be completed by expert judgements. Computing the stress scenarios Once we have estimated the parametersp,μ andσ, we can compute the stress scenarios using the Monte Carlo method. Nevertheless, we face an issue here, because the stress scenario is not unique to an investor category. Indeed, it depends on the liability structure of the fund. While the individual-based model is more realistic and relevant than the zero-inflated model, then it appears to be limited from a practical point of view. Nevertheless, it is useful to understand the importance of the liability structure on the redemption rate. We now consider an extension of the previous model by introducing correlations between the investors. We obtain the same expression of the redemption rate: However, the random variables (E 1 , . . . , E n , R 1 , . . . , R n ) are not necessarily independent. We discuss three different correlation patterns: 1. We can assume that E i and E j are correlated. This is the simplest and most understandable case. Indeed, we generally observe long periods with low redemption frequencies followed by short periods with high redemption frequencies, in particular when there is a crisis or a panic. We can assume that the redemption severities R i and R j are correlated. For example, it would mean that high (resp. low) redemptions are observed at the same time. Nevertheless, this severity correlation is different from the previous frequency correlation. Indeed, the severities are independent from the number of redemptions, implying that the severity correlation only concerns the unitholders that have already decided to redeem. 3 . We can assume that E i and R i are correlated. We notice that we can write the redemption rate for a given category as follows: where R i = E i ·R i is the individual redemption rate for the i th investor. The breakdown between the binary variable E i and the continuous variable R i helps us to model the "clumping-at-zero" pattern. But there is no reason that the value taken by the redemption severity R i depends whether E i takes the value 0 or 1, because R i is observed only if E i = 1. Finally, only the first two correlation patterns are relevant from a financial point of view, because the third correlation model has no statistical meaning. Nevertheless, it is obvious that the first correlation model is more appropriate because the second correlation model confuses low-severity and high-severity regimes. During a liquidity crisis, both the redemption frequency and the redemption severity increase (Coval and Stafford, 2007; Duarte and Eisenbach, 2013; Kacperczyk and Schnabl, 2013; Roncalli and Weisang, 2015a; Schmidt et al., 2016) . The first effect may be obtained by stressing the parameterp or by considering a high-frequency regime deduced from the first correlation model, but the second effect can only be obtained by stressing the parameterμ and cannot be explained by the second correlation model. Therefore, we only consider the first correlation pattern by modeling the random vector (E 1 , . . . , E n ) with a copula decomposition. We continue to assume that E i ∼ B (p) are identically distributed, but the dependence function of (E 1 , . . . , E n ) is given by the copula function C (u 1 , . . . , u n ). The individual-based model is then a special case of this copula-based model when the copula function is the product copula C ⊥ . In what follows, we consider the Clayton copula 25 : or the Normal copula: The Clayton parameter satisfies θ c ≥ 0 whereas the Normal parameter θ c lies in the range [−1, 1]. These two copula families are very interesting since they are positively ordered with respect to the concordance stochastic ordering. For the Clayton copula, we have: meaning that the product copula is reached when θ c = 0 and the upper Fréchet bound corresponds to the limiting case θ c → ∞. For the Normal copula, we restrict our analysis to θ c ∈ [0, 1] because there is no sense to obtain negative correlations. Therefore, we have: The Normal parameter θ c is easy to interpret because it corresponds to the Pearson linear correlation between two Gaussian random variables. The interpretation of the Clayton copula θ c is more tricky. Nevertheless, we can compute the associated Kendall's tau and Spearman's rho rank correlations 26 . Their expressions are given in Table 26 . Therefore, we can deduce the Pearson correlation ρ. The previous formulas can be used to map the copula parameter space into the Kendall, Spearman or Pearson correlation space. Some numeric values are given in Table 27 . For example, the Clayton copula θ c = 2 corresponds to a Kendall's tau of 50%, a Spearman's rho of 69% and a Pearson correlation of 70.7%. The following analyses will present the results with respect to the Pearson correlation, which is the most readable and known parameter. 25 We use the notations of Roncalli (2020, Chapter 11) . 26 For the Clayton copula, we calculate an approximation of the Spearman's rho: Remark 10 We denote the copula-based model by CM (n, ω,p,μ,σ, ρ) (or CM (n,p,μ,σ, ρ) when the vector of weights are equally-weighted). We have the following equivalence: IM (n, ω,p,μ,σ) = CM (n, ω,p,μ,σ, 0) The skewness effect In Appendix A.7.1 on page 87, we show that: Since C ⊥ ≺ C (θc) ≺ C + , we obtain the following bounds 27 : We notice that the probability to observe zero redemptions converges to zero only when the number n of unitholders tends to ∞ and the copula is the product copula. By assuming that the redemption frequencyp is equal to 10%, we obtain the results given in Figure 45 on page 102 and we verify the previous statistical property. In Figure 25 , we show the relationship between the Pearson correlation ρ and the probability Pr {R = 0}. As expected, it is an increasing function. We notice that the introduction of the correlation is very important to understand the empirical results we have calculated in Table 14 on page 36 and some unrealistic values we have obtained in Table 18 on page 40. For instance, the fact that Pr {R = 0} is equal to 54.39% for the retail category can only be explained by a significant frequency correlation since the number n of unitholders is high for this category. By construction, the frequency correlation modifies the probability distribution of the redemption frequency F , which is defined as the proportion of unitholders that redeem: F is a random variable, whose range is between 0 and 1. F depends on the frequency parameterp, the number n of unitholders and the copula function C (θc) (or the Pearson 27 Because we have: (1 −p) = (1 −p) n correlation ρ). When C (θc) is the product copula C ⊥ , the redemption events are independent and we obtain: F ∼ B (n,p) n because the sum of independent Bernoulli random variables is a binomial random variable. Therefore, we obtain the following approximation when n is sufficiently large: When the copula C (θc) corresponds to the upper Fréchet bound C + , the redemption frequency follows the Bernoulli distribution and does not depend on the number of unitholders: We have represented these two extreme cases in Figure 26 whenp = 20% and n = 1 000. We have also reported the probability distribution of F when the Pearson correlation of the copula function is equal to 25% and 50%. We notice that the skewness risk increases with the frequency correlation. Therefore, the parameter ρ will have a high impact on the stress testing results. In particular, when the frequency correlation is high, the risk is to observe a large proportion of redemptions even if the number of unitholders is large. In this case, the diversification effect across unitholders is limited. An illustration is provided in Figure 46 on page 102 that shows the probability to observe 100% of redemptions 28 when n is set to 20. The mean effect In Appendix A.7.3 on page 88, we show that the frequency correlation has no impact on the average redemption rate since we obtain the same expression as previously: Therefore, the redemption frequency changes the shape of the probability distribution of R , but not its mean. The volatility effect The volatility of the redemption rate is equal to: whereC (θc) is the survival copula associated to C (θc) . Since we have C ≺ C (θc) ≺ C + , we obtain the following inequalities: If we consider the equally-weighted case and assume that n tends to infinity, we obtain: This implies that the volatility risk is not equal to zero for an infinitely fine-grained liability structure if the frequency correlation is different from zero. The impact of the frequency correlation is illustrated in Figures 27 and 28 . We notice that the decrease of the volatility risk highly depends on the correlation parameter ρ. These figures confirm that the volatility risk is minimum when the frequency correlation is equal to zero. The consequence is that the frequency correlation is a key parameter when building stress testing scenarios. This is perfectly normal since ρ can been seen as a parameter that controls spillover effects and the magnitude of redemption contagion. All these results corroborate the previous intuition that the individual-based model without redemption correlation may be not appropriate for building a robust stress testing program. The shape effect The impact of the frequency correlation on the skewness and the volatility can then change dramatically the shape of the probability distribution of the redemption rate. In Figure 17 on page 39, we have already studied the histogram of the redemption rate in the casep = 50%,μ = 50% andσ = 10%. Let us reproduce the same exercise by assuming that the frequency correlation is equal to 50%. The results are given in Figure 29 . The shape of the probability distributions is completely different except in the case of a single unitholder 29 . To better illustrate the impact of the frequency correlation, we report in Figure 30 the histogram of the redemption rate by fixing n = 10. In the case of a perfect correlation of 100% and an equally-weighted liability structure, we obtain two different cases: 1. there is zero redemption with a probability 1 −p; 2. there are n redemptions with a probabilityp, and the redemption severity R is the average of the individual redemption severities: It follows that the probability distribution of the redemption rate is equal to: We retrieve the zero-inflated model ZI p,μ, n −1/2σ or the individual-based model with a single unitholder IM 1,p,μ, n −1/2σ . The only difference is the severity distribution G, whose variance is divided by a factor n. Spillover and contagion risks come then from the herd behavior of unitholders. Instead of having n different investors, we have a unique investor in the fund, because the decision to redeem by one investor induces the decision to redeem by all the other remaining investors. Correlation risk within the same investor category In order to illustrate that redemption frequencies are correlated, we build the time series of the frequency rate F t for a given category 30 : Other illustrations are provided in Appendix C on page 103. Figures 47, 48 and 49 correspond to the cases ρ = 25%, ρ = 75% and ρ = 90%. 30 We can use an equally-weighted scheme ω i,t = 1/n. Figure 29 : Histogram of the redemption rate in % with respect to the number n of unitholders (p = 50%,μ = 50%,σ = 10%, ρ = 50%) Figure 30 : Histogram of the redemption rate in % with respect to the frequency correlation (p = 50%,μ = 50%,σ = 10%, n = 10) where E i,t is the redemption indicator for the investor i at time t. Using the sample (F 1 , . . . , F T ), we compute the empirical mean F and the standard deviationσ (F ). Then, the copula parameter θ c can be calibrated by solving the following nonlinear equation 31 : The copula parameter θ c can be transformed into the Kendall, Spearman or Pearson correlation using the standard formulas given in Table 26 on page 52. For instance, if C (θc) is the Clayton copula, the Pearson correlation is equal to: An example is provided in Table 28 when the fund liability structure is equally-weighted and has 20 unitholders. For instance, if the empirical mean F and the standard deviation σ (F ) are equal to 25% and 20%, the calibrated Pearson correlation is equal to 44.5%. Remark 11 At first sight, calibrating the frequency correlation seems to be an easy task. However, it is very sensitive to the different parameters F ,σ (F ) and H (ω). Moreover, it depends on the copula specification. For instance, we obtain the results given in Table 49 on page 104 when the dependence function is the Normal copula. We observe that the Pearson correlations calibrated with the Clayton copula are different from those calibrated with the Normal copula. Remark 12 Another way to illustrate the frequency correlation is to split a given investor category into two subsamples S 1 and S 2 and calculate the time series of the redemption frequency for the two subsamples S k (k = 1, 2): Then, we can calculate the Pearson correlation ρ (F 1 , F 2 ) and calibrate the associated copula parameter θ c using Equation (58) on page 91. The correlation risk is present within a given investor category, but it may also concern two different investor categories. In order to distinguish them, we use the classical statistical jargon of inter-class and intraclass correlations. In Table 29 , we report the intra-class Spearman correlation 32 for four 31 See Equation (56) on page 90. 32 The correlations of retail/insurance and institutional/insurance for balanced funds and the correlations of retail/third-party distributor and retail/insurance for money market funds are not significant at the confidence level of 95%. investor categories (retail, third-party distributor, institutional and insurance) and four fund categories (balanced, bond, equity and money market). We observe a high inter-class correlation between retail investors and third-party distributors except for money market funds. We notice that equity and bond funds present very similar frequency correlations. On average, it is equal to 30%. For balanced and money market funds, we obtain lower figures less than 20%. These results are coherent with the academic research, since redemption runs and contagions in bond and equity funds have been extensively studied and illustrated (Lakonishok et al., 1992; Wermers, 1999; Sias, 2004; Wylie, 2005; Coval and Stafford, 2007; Shleifer and Vishny, 2011; Cai et al., 2019) . Remark 13 Another way to illustrate the intra-class correlation is to report the dependogram (or empirical copula) of redemption frequencies. An example is provided in Figure 31 for retail investors and third-party distributors. We observe that these dependogram does not correspond to the product copula 33 . The parameters of the copula-based model is made up by the parameters of the individualbased model (p,μ andσ) and the copula parameter θ c (or the associated frequency correlation). Once these parameters are estimated for a given investor/fund category, we transform theμ −σ parameterization into the a − b parameterization of the beta distribution and compute the risk measures M, Q (α), C (α) and S (T ) by using the following Monte Carlo algorithm: 1. we set k ←− 1; 2. we generate 34 (u 1 , . . . , u n ) ∼ C (θc) ; 3. we compute the redemption events (E 1 , . . . , E n ) such that: we simulate the redemption severities (R 1 , . . . , R n ) from the beta distribution 35 B (a, b); 5. we compute the redemption rate for the k th simulation iteration: if k is equal to n S , we return the simulated sample R (1) , . . . , R (n S ) , otherwise we set k ←− k + 1 and go back to step 2. Figure 32 shows the relationship between the correlation frequency 36 and C (99%) for different parameter sets when the liability structure has 20 equally-weighted unitholders. The impact of the correlation risk is not negligible in some cases. This is particularly true when the frequency correlation is close to 100%, but its impact is also significant when the frequency correlation is larger than 20%. On average, we observe that the risk measure C (99%) increases by 15%, 20% and 35% when the frequency correlation is respectively equal to 20%, 30% and 50% compared the independent case. The algorithm to simulate the copula-based model CM (n, ω,p,μ,σ, ρ) can be used to simulate the individual-based model IM (n, ω,p,μ,σ) by setting C (θc) = C ⊥ . This is equivalent to replace step 2 and simulate n independent uniform random numbers (u 1 , . . . , u n ). In the case of daily redemptions, the correlation risk only concerns the cross-correlation between investors for a given market day. When we consider fire sales or liquidity crisis, the one-day study period is not adapted and must be extended to a weekly or monthly basis. In this case, we may face time aggregation risk, meaning that redemption flows for the subsequent market days may depend on the current redemption flows. 34 Clayton and Normal copulas are easy to simulate using the method of transformation (Roncalli, 2020, page 803) . 35 Generally, the generation of beta random numbers is present in mathematical programming languages (Matlab, Python). Otherwise, we can use the method of rejection sampling (Roncalli, 2020, pages 886-887) . 36 It corresponds to the Pearson correlation of the Clayton copula. We recall that the total net assets at time t + 1 can be decomposed as follows: TNA (t + 1) = (1 + R (t + 1)) · TNA (t) + F + (t + 1) − F − (t + 1) By assuming that F + (t + 1) = 0, we obtain: This formula is valid on a daily basis. If we consider a period of n h market days (e.g. a weekly period), we have: Therefore, it is not obvious to decompose the difference TNA (t + n h ) − TNA (t) into a "performance" effect and a "redemption" effect since the two effects are related. Indeed, the mathematical definition of the n h -day redemption rate is: whereas the fund return over the period [t, t + n h ] is given by the compound formula: (1 + R (t + h)) − 1 Because of the cross-products (Brinson et al., 1991) , we cannot separate the two effects: TNA (t + n h ) = (1 + R (t; t + n h ) − R (t; t + n h )) · TNA (t) In the case where the performance effect is negligible -R (t + h) R (t + h), we have: We can then calculate the probability distribution of R (t, t + n h ) by the Monte Carlo method. A first solution is to consider that the redemption rates are time-independent. A second solution is to consider that redemption rates are auto-correlated: where ρ time is the autocorrelation parameter and ε (t) is a random variable such that R (t) ∈ [0, 1]. Such modeling is complex because of the specification of ε (t). However, this approach can be approximated by considering a time-series copula representation: whereF is the probability distribution of R (t) defined by the individual-based (or copulabased) model, C is the Normal copula, whose parameters are given by the Toeplitz cor- To calculate the probability distribution of R (t, t + n h ), we first simulate the individual-based (or copula-based) model in order to estimate the probability distributionF (x) of daily redemptions. Then, we generate the sample of the time-series (R (t + 1) , . . . , R (t + n n )) by using the method of the empirical quantile function (Roncalli, 2020, pages 806-809) . Finally, we calculate the redemption rate R (t, t + n h ) using Equation (28). An example is provided in Figure 33 when the correlation between investors is equal to zero 38 . We have also measured the impact of the autocorrelation value ρ time on the value-at-risk and the conditional value-at-risk. Results are given in Tables 30 and 31 for six different individual-based models IM (n,p,μ,σ). When the value of the risk measure is small, we notice that the impact of ρ time is high. For instance, when n = 500,p = 1%,μ = 25% andσ = 10%, the value-at-risk Q (99%) is equal to 1.9% in the independent case. This figure increases respectively by +9% and +19% when ρ time is equal to 25% and 50%. We also notice that the impact on the conditional value-at-risk is close to that on the value-at-risk. The compound approach defined by Equation (28) certainly overestimates stress scenarios. Indeed, we implicitly assume that the redemptions rates R (t + h) are identically distributed, meaning that there is no time effect on the individual redemption behaviour. However, we can think that an investor that redeems at time t + 1 will not redeem at time t + 2 and t + 3. In practice, we observe that redemptions of a given investor are mutually exclusive during a short period of time. This property is not verified by Equation (28). At time t + h, we notice IS (t + h) the set of investors that have redeemed some units before 37 For instance, in the case of a weekly period, the Toeplitz correlation matrix is equal to: The same example with a correlation of 50% between investors is given in Figure 52 on page 106. Figure 33 : Histogram of the weekly redemption rate in % with respect to the autocorrelation ρ time (p = 50%,μ = 50%,σ = 10%, ρ = 0%, n = 10) Table 30 : Impact of the autocorrelation ρ time on the value-at-risk Q (99%) npμσ ρ time 0% 25% 50% 75% 100% 10 000 0.1% 25% 10% 0.2% +6% +14% +24% +36% 500 1.0% 25% 10% 1.9% +9% +19% +33% +50% 50 2.0% 50% 10% 10.5% +12% +29% +49% +79% 100 5.0% 50% 30% 18.2% +8% +18% +29% +45% 10 20.0% 50% 30% 65.8% +6% +13% +21% +28% 10 50.0% 50% 30% 90.1% +2% +4% +6% +8% Table 31 : Impact of the autocorrelation ρ time on the conditional value-at-risk C (99%) npμσ ρ time 0% 25% 50% 75% 100% 10 000 0.1% 25% 10% 0.2% +6% +16% +27% +41% 500 1.0% 25% 10% 2.0% +9% +21% +37% +56% 50 2.0% 50% 10% 11.4% +13% +32% +54% +84% 100 5.0% 50% 30% 19.2% +9% +20% +32% +50% 10 20.0% 50% 30% 68.8% +6% +13% +21% +28% 10 50.0% 50% 30% 91.3% +2% +4% +6% +7% t + h. We have IS (t + 1) = {1, . . . , n}. The mutually exclusive property implies that 39 : It follows that: and : . Therefore, the redemption decisions taken in the recent past (e.g. two or three days ago) have an impact on the future redemptions for the next days. This is a limit of the compound approach. The solution would be to develop a comprehensive individual-based model, whose random variables are replaced by stochastic processes. Nevertheless, the complexity of such model is not worth it with respect to the large uncertainty of stress testing exercises. Herding risk is related to momentum trading. According to Grinblatt et al. (1995) , herding behavior corresponds to the situation where investors buy and sell the same securities at the same time. Herding risk happens during good and bad times, and is highly documented in economic research (Wermers, 1999; O'Neal, 2004; Ivković and Weisbenner, 2009; Ferreira et al., 2012; Lou, 2012; Cashman et al., 2014; Chen and Qin, 2017; Goldstein et al., 2017; Choi et al., 2019; Dötz and Weth, 2019) . However, we generally notice that sell herding may have more impact on asset prices than buy herding. Therefore, the sell-herding behavior risk may be associated to a price destabilizing or spillover effect. In the case of redemption risk, the spillover mechanism corresponds to two related effects: • A first spillover effect is that the unconditional probability of redemption is not equal to the conditional probability of the redemption given the returns of the fund during the recent past period: • A second spillover effect is that the unconditional probability of return is not equal to the conditional probability of the return given the redemptions of the fund during the recent past period: This implies that the transmission of a negative shock on the redemption rate R (t + 1) may also impact the redemption rates {R (t + 2) , R (t + 3) , . . .} because of the feedback Figure 34 : Spillover between fund redemptions and fund returns loop on the fund performance. An illustration is provided in Figure 34 . A large negative redemption R (t + 1) may induce a negative abnormal performance R (t + 1), and this negative performance may encourage the remaining investors of the fund to redeem, because negative returns accelerate redemption flows. This type of behavior is generally observed in the case of fire sales and less liquid markets. As explained in the introduction, an integrated model that combines liability risk and asset risk is too ambitious and too complex. Moreover, this means modeling the policy reaction function of other investors and asset managers. Nevertheless, if we want to take into account sell herding, spillover or fire sales, we must build an econometric model. For example, the simplest way is to consider the linear dynamic model: We obtain an AR(1) process: where φ = φ 1 φ 2 and u (t) = u 2 (t) + φ 2 u 1 (t − 1) is a white noise process. It follows that: Therefore, spillover scenarios can be estimated by applying a scaling factor to the initial shock 40 . In order to illustrate the time dependency between redemptions, we build the time series of the redemption rate R (j,k) (t), the redemption frequency F (j,k) (t) and the redemption severities R (j,k) (t) for each classification matrix cell (j, k), which is defined by a fund category FC (j) and an investor category IC (k) . For that, we calculate R (f,k) (t) the redemption rate of the fund f for the investor category IC (k) at time t. Then, we estimate the daily redemption rate R (j,k) (t) as the average of the redemption rates of all funds that belong to the fund category FC (j) : where 41 S (j,k) (t) = f : f ∈ FC (j) , TNA (f,k) (t) > 0 . We also estimate the daily redemption frequency as follows: 40 The previous analysis can be extended to more sophisticated process, e.g. VAR (p) processes. 41 We only consider funds which have unitholders that belong to the investor category IC (k) . This is equivalent to impose that the assets under management held by the investor category IC (k) are strictly positive: TNA (f,k) > 0. whereas the daily redemption severity is given by the following formula: where The computation of R (j,k) (t), F (j,k) (t) and R (j,k) (t) does make sense only if there is enough observations S (j,k) (t) and S (j,k) (t) at time t. This is why we focus on the most representative investor categories (retail, third-party distributor, institutional and insurance) and fund categories (balanced, bond, equity and money market). In Table 32 , we report the maximum between the autocorrelation ρ (R (t) , R (t − 1)) of order one and the autocorrelation ρ (R (t) , R (t − 2)) of order two. Moreover, we indicate with the symbol * * the matrix cells where the p-value of the autocorrelation is lower than 5%. Except for money market funds and the institutional/balanced matrix cell, redemptions are not significantly autocorrelated. If we consider redemption frequencies and severities, we observe more autocorrelation (see Tables 50 and 51 on page 106) . However, for bond and equity funds, the results show that the autocorrelation is significant and high for the redemption frequency, but low for the redemption severity. The last section of this article is dedicated to the factors that may explain a redemption stress. First, we investigate whether it is due to a redemption frequency shock or a redemption severity shock. Second, we study how market risk may explain extreme redemption rates, and we focus on three factors: stock returns, bond returns and volatility levels. We may wonder whether the time variation of redemption rates is explained by the time variation of redemption frequencies or redemption severities. Using the time series built in Section 4.3.4 on page 66, we consider three linear regression models: In the first model, we explain the redemption rate using the redemption frequency. In the second model, the explanatory variable is the redemption severity. Finally, the third model combines the two previous models. For each classification matrix cell (j, k), we have reported the centered coefficient of determination R 2 c in Tables 33, 34 and 35. If we consider the first linear regression model, we notice that R 2 c is greater than 50% only for the institutional/equity category. R 2 c takes a value around 35% for the retail/balanced, retail/bond and institutional/bond categories, otherwise it is less than 20%. Results for the second linear regression are better. This indicates that the redemption severity is a better explanatory variable than the redemption frequency. The only exception is the institutional/equity category. The combination of the two variables allows us to improve the explanatory power of the model, but we also notice that the redemption severity is the primary factor. The matrix cell with the highest R 2 c is retail/equity, whereas the matrix cell with the lowest R 2 c is institutional/equity. The scatter plot between R (t), F (t) and R (t) for these two extreme cases are reported in Figures 35 and 36 . For the retail/equity category, we verify that the redemption severity explains the redemption rate. For the institutional/equity category, the redemption severity is not able to explain the high values of the redemption rate. The previous results are very interesting since the redemption severity is the primary factor for explaining the redemption shocks. Therefore, a high variation of the redemption rate is generally due to an increase of the redemption severity. Nevertheless, there are some exceptions where stress scenarios are also explained by an increase in the redemption frequency. We have used the coefficient R 2 c to show the power explanation of the two variables F (t) and R (t) without considering the effect of the constant. For some matrix cells, we notice that the constant may be important (see Tables 52, 53 and 54 on page 106). What market risk factors matter in stress testing? 5.2.1 The flow-performance relationship Numerous academic research papers suggest that investor flows depend on past performance. According to Sirri and Tufano (1998) and Huang et al. (2007) , there is an asymmetry concerning the flow-performance relationship: equity mutual funds with good performance gain a lot of money inflows, while equity mutual funds with poor performance suffer smaller outflows. However, this asymmetry concerns relative performance. Indeed, according to Ivković and Weisbenner (2009) , "inflows are related only to relative performance" while "outflows are related only to absolute fund performance". Therefore, these authors suggest that investors sell the asset class when this one has a bad performance. In the case of corporate bonds, Goldstein et al. (2017) find that relative performance also matters in terms of explaining outflows. In order to better understand these results, we consider the following analytical model 42 : where R f (t) is the return of the fund f , R mkt (t) is the return of the market risk factor and R f (t) is the redemption rate of the fund f . ε (t) and η (t) are two independent white noise processes. Using the first equation, we can estimate the relative performance of the fund, which is measured by its alpha component α f (t). The second equation states that the redemption rate R f (t) of the fund depends on the past relative performance α f (t − 1) and the past absolute performance R f (t − 1). Then, we can test two assumptions: H 1 : δ f < 0 and H 2 : ϕ f < 0. Accepting H 1 implies that outflows depend on the relative performance, while accepting H 2 implies that outflows depend on the absolute performance. In both cases, the value of the coefficient is negative, because we expect that a negative performance will increase the redemption rate. The previous framework can be extended to take into account a more sophisticated model for determining the relative performance 43 α f (t) or to consider lagged variables (Bellando and Tran-Dieu, 2011; Ferreira et al., 2012; Lou, 2012; Cashman et al., 2014; Barber et al., 2016; Fricke and Fricke, 2017) . More generally, we have: Even if this type of flow-performance relationship is interesting to understand the investor behavior, it is however not adapted in the case of a stress testing program for two reasons. The first reason is that Equation (34) is calibrated using low frequency data, e.g. quarterly or monthly data. Therefore, the goal of Equation (34) is to describe long-term behavior of investors, whereas stress testing of liabilities concerns short-term periods. The second reason is the inadequacy of this approach with macro stress testing approaches developed by regulators and institutional bodies. If we consider stress testing programs developed in the banking sector (Roncalli, 2020, pages 893-922) , we distinguish historical, probabilistic and macroeconomic approaches. While the first two methods have been developed in the previous sections, we focus on the third method, which is the approach used by the regulators (Board of Governors of the Federal Reserve 42 See Arora et al. (2019) . 43 For instance, we can use the three-factor Fama-French model or the four-factor Carhart model. System, 2017; EBA, 2020a,b; ECB, 2019; Ong, 2014) . The macroeconomic approach consists in defining stress scenarios by a set of risk factors corresponding to some exogenous shocks. In this article, we focus on three market risk factors: • the performance of the bond market; • the performance of the stock market; • market volatility. Therefore, we assume that there is a linear relationship between the redemption rate and these factors: where F bond (t) and F stock (t) are the h-day total returns of the FTSE World Broad Investment-Grade Bond index and the MSCI World index, F vol (t) is the difference of the VIX index between t − h and t, and h is the time horizon. In Table 36 , we report the coefficient of determination R 2 c for the one-day time horizon. These figures are disappointing since the impact of the market risk factors are very low 44 . For instance, the highest R-squared is reached for the third-party distributor/money market category, but it is equal to 4.4%. If we consider a longer time horizon, results do not improve and we always have R 2 c 5% (see Tables 55 and 56 3 4.4 Remark 17 The previous results suggest that redemption rates do not depend on market risk factors on a short-term basis. However, fund managers generally have the feeling that redemption rates increase when there is a stress on market returns. Nevertheless, we know that returns are more or less independent from one day to another. Therefore, we consider another approach using market sentiment. For that, we compute the average redemption rate when the VIX index is above 30, and calculate its relative variation with respect to the entire period. Results are given in Table 37 . We observe an impact in particular for bond/equity funds and institutional/third-party distributor investors. Liquidity stress testing is a recent topic in asset management, which has given rise to numerous publications from regulators (AMF, 2017; BaFin, 2017; ESMA, 2019; FSB, 2017; IOSCO, 2015 IOSCO, , 2018 , investment management associations (AFG, 2015; EFAMA, 2020) and affiliated researchers from central banks and international bodies (Arora et al., 2019; Baranova et al., 2017; Bouveret, 2017; Fricke and Fricke, 2017; Gourdel et al., 2018) . On the academic side, few studies specifically concern liquidity stress testing in asset management 45 . Therefore, we observe a gap between general concepts and specific measurement models. As such, the purpose of our study is to propose several analytical approaches in order to implement LST practical programs. Besides the historical approach that considers non-parametric risk measures, we have developed a frequency-severity model that is useful when building parametric risk measures of liquidity stress testing. This statistical approach can be seen as a reduced-form model based on three parameters: the redemption frequency, the expected redemption severity and the redemption uncertainty. Like the historical approach, the frequency-severity model requires some expert judgements to correct some data biases. Nevertheless, both historical and analytical approaches are simple enough to verify properties of risk ordering coherency between fund and investor categories. We have also developed an individual-based behavioral model, which is an extension of the frequency-severity model. We have shown that redemption risk depends on the fund liability structure, and is related to the Herfindahl index of assets under management held by unitholders. Even if this model is hard to implement because it requires knowing the comprehensive liability structure, it allows us to justify liquidity stress testing based on the largest fund holders. Moreover, this model shows the importance of cross-correlation between unitholders of a same investor category, but also of several investor categories. Nevertheless, the individual-based behavioral model is flexible enough that it can easily take into account dependencies between investors by incorporating a copula model. Again, the issue with this extended individual-based behavioral model lies in the knowledge of the liability structure. The production of stress scenarios can be obtained by considering a risk measure applied to the redemption rate. For the historical approach, we can use a value-at-risk or a conditional value-at-risk figure, which is estimated with non-parametric statistical methods. For the frequency-severity and individual-based behavioral models, the estimation of the VaR or CVaR is based on analytical formulas. Moreover, these models may produce parametric stress scenarios for a given return time. Another issue concerns the choice of data between gross or net redemption rates for calibrating these stress scenarios. For some categories, net redemption rates may be used to proxy gross redemption rates, because they are very close in stress periods. However, we also demonstrate that it is better to use gross redemption rates for some investor or fund categories (e.g. retail investors or money market funds). The design of macro stress testing programs is more complicated than expected. Since the flow-performance relationship is extensively documented by academic research, it is valid at low frequencies, typically on a quarterly or annual basis. In this case, we may observe inflows towards the best fund managers. However, this relationship mainly concerns relative performance, whereas macro stress testing programs deal with absolute performance. Indeed, relative performance is a key parameter when we want to analyze the idiosyncratic liability liquidity risk at the fund level. Nevertheless, the liquidity risk in asset management primarily involves systemic periods of liquidity shortage that impact a given asset class. Our empirical results are mixed since drawing a relationship between redemption rates and market risk factors in stress periods is not obvious because there are lead/lag effects and liquidity stress periods never look the same. For instance, the redemption stress scenario on money market funds during the covid-19 crisis and the first quarter of 2020 is very different from the redemption stress scenario during the Lehman Brothers' bankruptcy in September and October 2008. Indeed, we observe a significant lag of one/two months in the case of the covid-19 crisis. In a similar way, the liquidity stress transmission to equity funds has not been immediate and has been delayed by several weeks. The current interest in liquidity stress testing is related to the Financial Stability Board's tasks on systemic risk (FSB, 2010 (FSB, , 2015 and shadow banking supervision (FSB, 2017 (FSB, , 2018 . As explained by Blanqué and Mortier (2019b) , "regulation of asset managers has been lagging behind that of banks since the global financial crisis". The implementation of the liquidity coverage ratio (LCR) and the net stable funding ratio (NSFR), the use of liquidity and high-quality liquid assets (HQLA) buffers and the definition of regulatory monitoring tools date back to 2010 for the banking industry 46 (BCBS, 2010 (BCBS, , 2013 . The regulatory framework on liquidity stress testing proposed by ESMA (2019) is an important step for the development of liquidity measurement in the asset management industry. In this paper, we develop an analytical framework and give some answers. However, it is still early days and much remains to be done. A Mathematical results A.1 Granularity and the X-statistic We consider n funds whose redemption rate is equal to p. The assets under management of each fund are set to $1. The maximum redemption rate of n funds is equal to the mathematical expectation of n Bernoulli random variables: whereas the redemption rate of the sum of n funds is equal to the expected frequency of a Binomial random variable: In Table 38 , we report the value taken by the ratio p (max) /p (sum). For example, this ratio is equal to 3.71 if p = 5% and n = 4. To understand this ratio, we can consider a large fund whose redemption probability is p. This fund is split into n funds of the same size. The ratio indicates the multiplication factor to obtain the maximum of the redemption rates among the n funds. A zero-inflated random variable Z can be written as the product of a Bernoulli random variables X ∼ B (p) and a positive random variable Y : Let µ m (Z) for the m-th moment of Z. Using the previous relationship, we deduce that: because X and Y are independent by definition, and X m = X, implying that X m follows a Bernoulli distribution B (p). From Equation (36), we can compute the m-th centered moment µ m (Z). For that, we recall that: We deduce the expression of the second moment: For the third moment, we have: where γ 1 is the skewness coefficient. For the fourth moment, it follows that: = µ 4 + 4 γ 1 µ 3/2 2 + 3µ 2 µ 1 + µ 3 1 µ 1 − 6 µ 2 + µ 2 1 µ 2 1 + 3µ 4 1 = µ 4 + 4γ 1 µ 3/2 2 µ 1 + 12µ 2 µ 2 1 + 4µ 4 1 − 6µ 2 µ 2 1 − 6µ 4 1 + 3µ 4 1 = µ 4 + 4γ 1 µ 3/2 2 µ 1 + 6µ 2 µ 2 1 + µ 4 1 = (γ 2 + 3) µ 2 2 + 4γ 1 µ 3/2 2 µ 1 + 6µ 2 µ 2 1 + µ 4 1 where γ 2 is the excess kurtosis coefficient. We can then compute the moments of Z. For the mean, we have: We deduce that the variance of Z is equal to: For the third moment, we have: It follows that the skewness coefficient is equal to: where: For the fourth moment, we have: We deduce that the excess kurtosis coefficient is equal to: where: We can deduce the following properties: 1. The skewness of Z is equal to zero if and only if: (a) the skewness of Y is equal to zero and the frequency probability p is equal to one; (b) the frequency probability p is equal to zero, meaning that Z is always equal to zero. (a) the kurtosis of Y is equal to 3 and the frequency probability p is equal to one; (b) the frequency probability p is equal to zero, meaning that Z is always equal to zero. In other cases, the skewness and excess kurtosis coefficients of Z are different from zero even if the random variable Y is not skewed and has not fat tails. The previous results seem to be contradictory with the properties given in Equation (17) on page 26. In fact, the limit case p → 0 + is not equal to p = 0, because there is a singularity at the point p = 0. We assume that Y ∼ B (a, b). Since we have: For the second moment, we have: and: This formula has been already found by Ospina and Ferrari (2010) . The skewness and excess kurtosis coefficients of the beta distribution are equal to: √ ab and: We plug these different expressions into the general formulas 47 to obtain γ 1 (Z) and γ 2 (Z). We consider a sample {x 1 , . . . , x n } of n observations, and we assume that X follows a zero-inflated model, whose frequency and probability distributions are p and G (x; θ). The log-likelihood of the i th observation is equal to: The formulas are not reported here because they don't have a lot of interest. We deduce that the log-likelihood function is equal to: where n 0 is the number of observations x i that are equal to zero. The maximum likelihood estimator p,θ is defined as follows: p,θ = arg max (p, θ) and satisfies the first-order conditions: We deduce that: The concentrated log-likelihood function becomes: (p, θ) = n 0 ln n 0 + (n − n 0 ) ln (n − n 0 ) − n ln n + xi>0 ln g (x i ) Therefore, the ML estimatorθ corresponds to the ML estimator of θ when considering only the observations x i that are strictly positive: Remark 19 In the case of the zero-inflated beta model, we have θ = (a, b) and: We define the random variableZ as the sum of products of two random variables: whereX i ∼ B (p) andỸ i are iid random variables. Moreover, we assume that ω i > 0 and n i=1 ω i = 1. This case corresponds to the situation where no client redeems: A. 4 First moment For the mean, we have: We deduce that: Second moment Since we have E X 2 i =p and E Ỹ 2 i = µ 2 Ỹ , it follows that: We notice that: Therefore, the variance ofZ is equal to: Remark 20 In the equally-weighted case, we obtain: Application to the beta severity distribution If we assume thatỸ i ∼ B ã,b , we have: In order to calibrate the probability p, we match the redemption probability Pr {R > 0}. Using the results in Appendix A. 4 .1 on page 83, we obtain: For the first moment, we have: We deduce that: For the second moment, we have: It follows that: where H (ω) = n i=1 ω 2 i is the Herfindahl index. If we consider the equally-weighted case ω i = n −1 , we have H (ω) = n −1 and: Let π k be a probability distribution, meaning that π k ≥ 0 and n k=1 π k = 1. The Herfindahl index is equal to: 0 ≤ min π k = π 1:n ≤ π 2:n ≤ · · · ≤ π k:n ≤ π k+1:n ≤ · · · ≤ π n:n = max π k We have: n−k+1:n + n k=m+1 π 2 n−k+1:n where k = 1 : m denotes the largest contributions that are known, meaning that we don't know the values taken by {π 1:n , . . . , π n−m:n }. Since we have π n−k:n ≤ π n−k+1:n , we deduce that: n k=m+1 π 2 n−k+1:n ≤ 1 − m k=1 π n−k+1:n π n−m+1:n π 2 n−m+1:n = 1 − m k=1 π n−k+1:n π n−m+1:n and 48 : π 2 n−k+1:n + 1 − m k=1 π n−k+1:n π n−m+1:n An example is given in Table 39 . The Herfindahl index is equal to 17.96%. Using the first three largest values, we obtain an estimate of 20.50%. We define the random variableZ as previously: ω iXiỸi 48 We verify that H + n = H. whereỸ i are iid random variables. We assume thatX i ∼ B (p) are identically distributed, but not independent. We note C (u 1 , . . . , u n ) the copula function of the random vector X 1 , . . . ,X n and B (x) the cumulative distribution function of the Bernoulli random variable B (p). This means that B (0) = 1 −p and B (1) = 1. In practice we use the Clayton copula: or the Normal copula 49 : The Clayton parameter satisfies θ c ≥ 0 whereas the Normal parameter θ c lies in the range [−1, 1]. Moreover, we notice that the expressions of the bivariate copula functions are: We consider the bivariate case. The probability mass function is described by the following contingency table:X 2 = 0X 1 = 1 X 1 = 0 π 0,0 π 0,1 π 0 = 1 −p X 1 = 1 π 1,0 π 1,1 π 1 =p π 0 = 1 −p π 1 =p 1 Since we have Pr X 1 ≤ u 1 ,X 2 ≤ u 2 = C (θc) (B (u 1 ) , B (u 2 )), we deduce that: (1)) = C (θc) (1, 1) = 1 and:X In the case where C (θc) = C ⊥ ,X 1 andX 2 are independent, we retrieve the results obtained for the individual-based model:X 2 = 0X 1 = 1 X 1 = 0 (1 −p) (1 −p)p 1 −p X 1 = 1 (1 −p)pp 2p 1 −pp 1 49 The Normal copula depends on the correlation matrix Σ. Here, we assume a uniform redemption correlation, implying that Σ is the constant correlation matrix Cn (θc) where θc is the pairwise correlation. because C ⊥ (u 1 , u 2 ) = u 1 u 2 . In the case where C (θc) = C + ,X 1 andX 2 are perfectly dependent and we obtain the following contingency table: because C + (u 1 , u 2 ) = min (u 1 , u 2 ). The contingency tables (51) and (52) represent the two extremes cases. If we use a radially symmetric copula (Nelsen, 2006) such that: the contingency table (50) becomes:X 2 = 0X 1 = 1 X 1 = 0 1 − 2p + C (θc) (p,p)p − C (θc) (p,p) 1 −p X 1 = 1 p − C (θc) (p,p) C (θc) (p,p)p 1 −pp 1 In the general case, we obtain a similar contingency table by replacing the copula function C (θc) (u 1 , u 2 ) by its corresponding survival functionC (θc) (u 1 , u 2 ) because we have (Nelsen, 2006) :C In the case where C (θc) = C ⊥ , we retrieve the result Pr Z = 0 = (1 −p) n . In the case where C (θc) = C + , we obtain Pr Z = 0 = 1 −p. First moment For the mean, we have: We deduce that: and: We deduce that: µ 2 (F ) =pH (ω) +C (θc) (p,p) (1 − H (ω)) −p 2 We notice that the expected value and the volatility of the redemption frequency are related in the following way: We consider two redemption frequencies F 1 and F 2 . The redemption frequency F k is associated to the liability structure (ω k,1 , . . . , ω k,n k ) and corresponds to an investor category, whose redemption probability isp k and frequency correlation is characterized by the copula function C (θ k ) (k = 1, 2). We also assume that the redemption correlation between the two investor categories is defined by the copula function C (θ12) . It follows that we have three copula functions: • C (θ1) is the copula function that defines the frequency correlation between the investors of the first category; • C (θ2) is the copula function that defines the frequency correlation between the investors of the second category; • C (θ12) is the copula function that defines the frequency correlation between the investors of the first category and those of the second category. In the case where the two categories are the same, we have C (θ1) = C (θ2) = C (θ12) = C (θc) . To compute the covariance between F 1 and F 2 , we calculate the mathematical expectation of the cross product: j=1 ω 1,i ω 2,jX1,iX2,j   = E X 1,iX2,j   n1 i=1 n2 j=1 ω 1,i ω 2,j   =C (θ12) (p 1 ,p 2 ) because n1 i=1 n2 j=1 ω 1,i ω 2,j = 1. We deduce the expression of the Pearson correlation: where: µ 2 (F k ) =p k (H (ω k ) −p k ) +C (θ k ) (p k ,p k ) (1 − H (ω k )) k = 1, 2 Remark 24 The Pearson correlation ρ (F 1 , F 2 ) is equal to zero if only if 50 C (θ k ) is the product copula C ⊥ . In the case where the two investor categories are the same and the liability structures are equally-weighted, we havep 1 =p 2 =p and C (θ1) = C (θ2) = C (θ12) = C (θc) , and we obtain: where: µ 2 (F k ) =C (θc) (p,p) −p 2 +p −C (θc) (p,p) n k k = 1, 2 The limiting case n k → ∞ is equal to ρ (F 1 , F 2 ) = 1. This is normal since F 1 and F 2 converges top when the liability structure is infinitely fine-grained. B Data 50 We recall that C (θ k ) is the Clayton or the Normal copula. In the general case, this property does not hold. (1) = balanced, (2) = bond, (3) = enhanced treasury, (4) = equity, (5) = money market, (6) = other, (7) = structured, (8) = total (1) = balanced, (2) = bond, (3) = enhanced treasury, (4) = equity, (5) = money market, (6) = other, (7) = structured, (8) = total (1) = balanced, (2) = bond, (3) = enhanced treasury, (4) = equity, (5) = money market, (6) = other, (7) = structured, (8) = total (1) = balanced, (2) = bond, (3) = enhanced treasury, (4) = equity, (5) = money market, (6) = other, (7) = structured, (8) = total (1) = balanced, (2) = bond, (3) = enhanced treasury, (4) = equity, (5) = money market, (6) = other, (7) = structured, (8) = total (1) = balanced, (2) = bond, (3) = enhanced treasury, (4) = equity, (5) = money market, (6) = other, (7) = structured, (8) = total Figure 52 : Histogram of the weekly redemption rate in % with respect to the autocorrelation ρ time (p = 50%,μ = 50%,σ = 10%, ρ = 50%, n = 10) Bond Funds and Fixed-Income Market Liquidity: A Stress-Testing Approach, Bank of Canada Technical Code of Conduct on the Use of Stress Tests The Use of Stress Tests as Part of Risk Management Simulating Stress Across the Financial System: The Resilience of Corporate Bond Markets and the Role of Investment Funds Which Factors Matter to Investors? Evidence from Mutual Fund Flows Amendment to the Capital Accord to Incorporate Market Risks A New Capital Adequacy Framework, First Consultative Paper on Basel II The New Basel Capital Accord, Second Consultative Paper on Basel II Basel III: A Global Regulatory Framework for More Resilient Banks and Banking Systems Basel III: The Liquidity Coverage Ratio and Liquidity Risk Monitoring Tools Interest Rate Risk in the Banking Book Stress Testing Principles, Consultative Document Fund Flow/Performance Relationship -A Case Study of French Mutual Funds How Investors Should Deal with the Liquidity Dilemna, Amundi CIO Insights Liquidity Dilemna Needs A Regulatory Response, Amundi Investment Talks Supervisory Scenarios for Annual Stress Tests Required under the Dodd-Frank Act Stress Testing Rules and the Capital Plan Rule Liquidity Stress Tests for Investment Funds: A Practical Guide, IMF Working Paper Vulnerable Asset Management? The Case of Mutual Funds Investor Flows and Fragility in Corporate Bond Funds Investments Funds Under Stress, ECB Working Paper Macro Stress Tests: What do they mean for the Markets and for the Asset Management Industry? Momentum Investment Strategies, Portfolio Performance, and Herding: A Study of Mutual Fund Behavior Participation Costs and the Sensitivity of Fund Flows to Past Performance Liquidity Management Tools in Collective Investment Schemes Recommendations for Liquidity Risk Management for Collective Investment Schemes Individual Investor Mutual Fund Flows How Safe are Money Market Funds? The Impact of Institutional Trading on Stock Prices A Flow-based Explanation for Return Predictability Modeling Nonnegative Data with Clumping at Zero: A Survey An Introduction to Copulas Purchase and Redemption Patterns of US Equity Mutual Funds A Guide to IMF Stress Testing: Methods and Models Inflated Beta Distributions, Statistical papers Liquidity Black Holes: Understanding, Quantifying and Managing Financial Handbook of Financial Risk Management Asset Management and Systemic Risk, SSRN Response to FSB-IOSCO Second Consultative Document, Assessment Methodologies for Identifying Non-Bank Non-Insurer Global Systemically Important Financial Institutions Runs on Money Market Mutual Funds Open-End Fund Liquidity Risk Management Programs; Swing Pricing; Re-Opening of Comment Period for Investment Company Reporting Modernization Release, Proposed Rule Investment Company Liquidity Disclosure, Proposed Rule Investment Company Liquidity Disclosure, Final Rule Fire Sales in Finance and Macroeconomics Institutional Herding Costly Search and Mutual Fund Flows H2O, Woodford and GAM Crises Highlight Liquidity Risk Mutual Fund Herding and the Impact on Stock Prices Fund Manager Herding: A Test of the Accuracy of Empirical Results using UK Data 1.2 The effect of subscriptions and redemptions 3.2 Net flow, net redemption and gross redemption rates The frequency-severity modeling approach 24 25 3.1.2 Statistical risk measures of the zero-inflated model 2.2 Stress scenarios based on the p − µ − σ parameterization Behavioral modeling 38 The individual-based model 44 Approximation of the probability distributionF (x | ω) 48 Using collective and mutual funds 57 Correlation risk within the same investor category Computing the stress scenarios 61 4.3.1 Analysis of non-daily redemptions Factor-based liquidity stress testing 67 Where does the stress come from? 2 What market risk factors matter in stress testing? Statistical moments of zero-inflated probability distribution Maximum likelihood of the zero-inflated model A.4 Statistical properties of the individual-based model 83 Application to the beta severity distribution Moment matching between the zero-inflated model and the individual-based model Upper bound of the Herfindahl index under partial information 86 A.7.1 Joint probability of twoX i 's Statistical moments of the redemption frequency Pearson correlation between two redemption frequencies Liquidity Stress Testing in Asset Management