Capa 85 ING.indd - Serasa Experian
Transcrição
Capa 85 ING.indd - Serasa Experian
Proteja A Serasa Experian oferece soluções sob medida para o seu negócio. São produtos e serviços exclusivos, que vão desde consultas de crédito até ferramentas e softwares para gerir melhor sua empresa. Tudo muito prático, rápido e rentável. E o melhor: cabe no seu orçamento. Entre em contato e prepare-se para se destacar da concorrência. Para saber mais, acesse serasaexperian.com.br ou ligue para 0800 773 7728 SoluçõeS eSPeCífiCaS Para Cada etaPa do Seu negóCio Busque novos clientes Venda com segurança Monitore empresas, clientes e sócios Cobre devedores Compre só os dados de que precisa 3 06 Homogeneous Risk Groups: A Proposal Using Cluster Analysis Fabio Augusto Scalet Medina Edson Luiz de Carvalho Barbosa The article investigates the applicability of Cluster Analysis to segregating a portfolio of retail exposures into homogeneous risk groups, according to the rules set forth in Central Bank of Brazil Circular Letter No. 3.648/2013. The study concludes that Cluster Analysis is a viable option. 21 Survival Analysis in Credit Models: a View of Covariates Treatment Melissa Sousa Tiago Lima Wanderson Rocha The study proposes adjusting a credit model based on survival analysis, with a main focus on the treatment of predictive variables and the creation of homogeneous groups. The method’s advantage over today’s logit regression lies in its ability to predict defaulters over different horizons. 31 Spatial Correlation of Corporate Defaults Guilherme B. Fernandes Rinaldo Artes Local economic effects may provide information on the default rate of individuals or business firms that live or operate in the relevant regions, and spatial analysis techniques enable measurement. The study uses the Moran’s I to evaluate the spatial correlation of defaults among São Paulo State business firms. 39 Infrastructure Credit and Financing in Brazil Frederico A. Turolla Márcio F. Gabrielli Igor J. C. Gondim The financial and capital markets fund infrastructure projects in many ways, with both debt and equity. Financing environments include corporate finance, which creates liabilities directly on the service operators’ balance sheets, and project finance, which is associated with Special Purpose Societies (“Sociedades de Propósito Específico” – SPE). 4 From the Editors The articles “Homogeneous Risk Groups: A proposal using Cluster Analysis” and “Survival Analysis in Credito Models: a View of Covariates Treatment”, selected for this issue of Tecnologia de Crédito review, address a methodology capable of predicting default among homogeneous groups. That is, risk groups in which a set of exposures prevails with shared characteristics for the purposes of credit risk evaluation and quantification. In the study “Homogeneus Risk Groups” by Statistics experts Fabio Augusto Scalet Medina and Edson Luiz de Carvalho Barbosa, the groups formed are well distributed across the number of contracts. They were shown to be consistent through analysis of characteristics vectors and quite distinctive in terms of percentage of default, one of the most important points to support the hypothesis of using cluster analysis to place retail exposures into homogeneous risk groups. According to the authors, a subject for future studies lies in developing a methodology to allocate exposures with observation vectors not present in the development sample, or classes to be created over time within one of the resulting homogeneous groups. In their turn, statisticians Melissa Sousa, Tiago Lima and Wanderson Rocha, who specialize in credit risk modeling, present a study - Survival Analysis in Credit Models - Analysis in Credito Mdels that proposes using survival analysis to adjust a credit model, with a main focus on the treatment of predictive variables and particular attention to the creation of homogeneous groups. The study’s highlight is a concern with the relationship between output and predictive variables, and the authors point out the presence of two main lines of thinking: occurrence of default and time to default. They use the Cox non-parametric model to perform the adjustment in such a manner as to align time to default and rate of default. The Logrank test was used at two stages in the study: treatment of the predictive variables to ensure classes with distinctive survival curves, and the creation of homogeneous groups that, in addition to sporting different rates of default over time, also show distinctive survival curves. This issue also contains an important article on modeling – “Spatial Correlation of Corporate Defaults”. Its authors, statisticians Guilherme Fernandes and Prof. Rinaldo Artes, note that, during development of a credit model, a spatial correlation may occur among small and medium-sized business firms due to the non-observation of local economic activity factors. The level of the correlation depends on the definition of the regions to be an- 5 alyzed, and this study finds that spatial correlat Analysis in Credito Mdels ion as measured by Moran’s I is weak when only 19 São Paulo State regions (as defined by the first two Postal Code numerals) are analyzed. However, when the area is defined based on sector (first four Postal Code numerals), Moran’s I shows the presence of a reasonably high spatial correlation. If development of a credit risk model fails to take this information into consideration, the model’s parameters may be biased as a result of the endogeneity created by omitted variables. Professors Frederico Araujo Turolla and Márcio Fernandes Gabrielli, and consultant Igor J. C. Gondim are the authors of “Infrastructure Credit and Financing in Brazil”. They discuss certain relevant challenges to infrastructure financing. That is, they point out the importance of costto-benefit analysis in taking large fiscal risks and of whether public and private financing modes are substitute or complementary goods. They also inquire whether a higher share of credit can be achieved relative to equity in infrastructure projects. They emphasize the need to re-discuss the institutional environment where industry risks are more relevant, particularly as concerns stability and the introduction of competition. The challenges are complex and inter-related, but offer fundamental answers for Brazil’s development. CREDIT TECHNOLOGY YEAR XIII Trimonthly published by Serasa Experian Nº 85 ISSN 2177-6032 President - Brazil Desktop Publishing Ricardo Loureiro Eric Miranda e Gerson Lezak Business Units Presidents/Superintendents Illustration Igor Ramos Rocha, Juliana Azuma, Marcelo Kekligian, Maria Eric Miranda e Gerson Lezak Zanforlin and Steven Wagner Translation Directors Allan Hastings Amador Rodriguez, Guilherme Cavalieri, Laércio Oliveira Pinto, Correspondência Lisias Lauretti, Paulo Melo, Silvânio Covas and Valdemir Bertolo Serasa Experian - Comunicação & Branding Responsible Editor Al. dos Quinimuras, 187 - CEP 04068-900 - São Paulo - SP Rosina I. M. D’Angina (MTb 8251) www.serasaexperian.com.br Assistant Editor [email protected] Nancy Galvão The concepts issued in the signed articles are the responsibility Graphic Design of the authors, which do not necessarily express the point of view Luis Barbuda of Serasa Experian and the Editorial Council. Cover Total or partial reproduction of the articles hereby published Eric Miranda e Gerson Lezak is strictly forbidden. 6 Homogeneous Risk Groups: A Proposal Using Cluster Analysis Fabio Augusto Scalet Medina Edson Luiz de Carvalho Barbosa 7 Introdução Abstract The purpose Nas últimas décadas, of this paper a evolução was to determine das tecnologias the applicade informação bility of Cluster e de transmissão Analysis to segregate de dadosaeportfolio a criação of da retail internet expomarcaram sures into homogeneous uma nova fase risk do processo groups according de globalização to the rules da ecoof nomia. Nesse Central Bank ofnovo Brazil cenário Circular surge, Lettercomo No. 3.648/2013, grande inovação which sets no mundo econômico, minimum requirements o comércio for the calculation eletrônico,ofpossibilitando the element assoa realização ciated with de credit operações risk exposures comerciais subject sem ato defi required nição exata capital docallocal de origem culation undereinternal destino.credit risk rating systems (IRB approaches) (RWACIRB ). The study’s main goal was achieved, proving that Cluster Analysis is aComércio viable option to segregate a retail porPalavras-chave: Eletrônico, Gestão da Frautfolio into homogeneous risk groups de, Informações de Bureau. Keywords: Basle II, Retail, Homogeneous risk groups, Cluster Analysis. Introduction In order to standardize supervision and discuss ways to strengthen the international banking system’s security, G-10 banking supervisors (Central Banks) established in 1974 the Basel Committee on Banking Supervision, headquartered in the Bank for International Settlements, in Basel, Switzerland. As a means to manage the risk of third-party funds leveraging on the part of Financial Institutions, in June 1988 the Basel Committee issued the International Convergence of Capital Measurement and Capital Standards, also known as Basel Agreement, or Basel I, whose main purpose was to require financial institutions to maintain sufficient capital to cover potential asset value losses and thereby ensure solvency. Several changes occurred in the financial market since Basel I. Among them, Jorion (2010) points out the inter- ruption of the economic and monetary unification path in Europe as a result of the European Monetary System’s crisis in September 1992; in the aftermath of the securities crisis of 1994, the Federal Reserve Bank (FED), after keeping interest rates low for three years, began a series of six consecutive hikes, causing $1.5 trillion in global capital to disappear; the Asian crisis of 1997 and the Russian default of 1998, which is regarded as the trigger of a global financial crisis. These changes led the Basel Committee to review the 1988 framework and release the New Capital Accord, or Basel II, in June 2004. According to Carneiro, Vivan and Krause (2005), Basel II proposes a new capital requirements framework based on three pillars: pillar I addresses capital requirements based on operational, market and credit risks; pillar II reinforces the ability of banking su- 8 pervisors to evaluate and adapt capital requirements to the individual circumstances of financial institutions; and the third pillar casts transparency and disclosure in an important and relevant role for the development of market discipline. Pillars I and II contain the calculation of Reference Equity (RE), which represents the capital a financial institution must have to face the risk of extreme losses in their exposures portfolios. Resolution No. 4.193, dated March 1st, 2013, governs the determination of minimum requirements for Reference Equity (RE), Level I Capital, and Core Capital, and institutes Additional Core Capital. According to the Resolution, for the purposes of minimum requirements and Additional Core Capital, the amount of Risk-Weighted Assets (RWA) must be determined as the sum total of the following (Equation 1), where RWACPAD and RWAconcern credit risk. CIRB According to Assaf Neto (2011, p. 136), credit risk “is the possibility of a financial institution not collecting on the amounts (principal and interest) promised by the securities it has in its receivables portfolio.” Jorion (2010, p.15), in his turn, reports that “credit risk arises when the counterparts will not or cannot honor their contractual obligations.” Basel II admits two alternative approaches to measuring asset risk for credit-risk purposes: the standard approach, represented by RWACPAD and the internal risk ratings-based approach (IRB), represented by RWACIRB. According to the standard approach, which is governed in Brazil by Circular Letter No. 3.644, dated March 4th, 2013, RWACPAD must equal the sum of the products of the exposures multiplied by the respective Risk-Weighting Factors (RWF), where RWF varies according to the risk assigned to exposure types. IRB approaches, in their turn, are based on estimates of the following risk elements: • Expected Losses - EL; • Unexpected Losses - UL; • Probability of Default - PD; • Loss Given Default - LGD; • Exposure at Default - EAD; • Maturity - M. Central Bank of Brazil (BACEN) Circular Letter No. 3.648, dated March 4th, 2013, establishes the minimum requirements for calculating credit risk exposures subject to required capital calculation based on internal credit risk rating systems (IRB approaches) (RWACIRB) According to Circular Letter No. 3.648/2013, exposures subject to the IRB approach for the purposes of regulatory capital determination must be segmented as follows: • Sovereign Entities; • Financial Institutions; • Retail; • Stockholdings; • Wholesale. RWA = RWACPAD + RWACIRB + RWAMPAD + RWACINT + RWAOPAD + RWAOAMA Credit Risk Market Risk Operational Risk (1) 9 The retail category includes: a) exposures to individuals and legal entities with gross annual sales under R$3.6 million, managed non-individually by means of homogeneous risk groups, and in the shape of typically retail-oriented financial instruments; and b) exposures from loans and financings to individuals with home equity for collateral; Chapter VI of the Circular letter in question addresses the retail category and provides the following definition in Section One, Article 44: “For exposures in the “Retail” category, credit risk-based exposure rating systems must enable associating each exposure to a certain homogeneous risk group (...).” Paragraph 1 of the same section reads: “A homogeneous risk group is defined as the group of retail exposures with common characteristics for the purposes of credit risk evaluation and quantification (...).” Still according to Article 44, the association of exposures to a certain homogeneous risk group must be based on the following criteria: • Risk characteristics of the obligor or counterpart; • Risk characteristics of the exposure, including product type and collateral, among others; and • Lateness in the operations associated with the exposures. Among the appropriate statistical techniques for the classifying a group of exposures according to homogeneous groups, the option of cluster analysis stands out. The technique can be found in several practical expe- riments in various domains. One example of the application of cluster analysis in the medical area can be found in Rapeli e Botega (2005), where cluster analysis is applied to identify the existence of different groups among individuals who attempted suicide with more severe clinical or surgical impact and who needed to be admitted to the UNICAMP Hospital das Clínicas. The study identified three groups of patients with different profiles in terms of the method used in their attempts. In the financial arena, Sanvicente and Minardi (1999) developed a risk rating system for Brazilian listed corporations using cluster analysis. The rating system was built using only accounting data from 301 publicly listed companies. After they were grouped, each cluster was assigned a rating. Given the recent publication of Circular Letter No. 3.648, there is little literature on the segregation of retail exposures into homogeneous risk groups. No application using cluster analysis was found in the Brazilian literature. Therefore, this paper will introduce a proposal for categorizing retail exposures according to homogeneous risk groups based on cluster analysis. By the end of the study, we expect to demonstrate that cluster analysis is applicable to the purpose. 2. Literature Review 2.1 Cluster Analysis According to Barroso and Artes (2003), cluster analysis is the name 10 of a series of techniques used to identify behavior patterns in data bans by means of the formation of homogeneous groups. Johnson and Wichern (2007) address cluster analysis as an important exploratory technique, as, by studying a natural group structure, it enables evaluating data dimensionality, identifying outliers and raising hypotheses relative to the objects’ structure (associations). Mingoti (2005) reports that cluster analysis aims to segregate the sample’s or population’s elements into groups so that elements belonging to the same group are similar to one another and different from the measured variables (characteristics) and elements in other groups. 2.1.1 Resemblance metrics Resemblance metrics play a very important role in clustering algorithms. They define the criteria to determine whether two points are close and, therefore, may or may not belong in the same group. There are two kinds of resemblance metrics: • Similarity metrics (the bigger the value, the greater the similarity between objects); • Dissimilarity metrics (the bigger the value, the greater the difference between objects). According to Landim (2000), if the data to be analyzed are a mixture of continuous values and binary measurements, the Gower general similarity coefficient can be applied, as it is a metric coefficient. According to the Help function of the SAS statistics package, version 9.3, the Gower similarity coefficient accepts all kinds of variables, including interval, ratio, ordinal, nominal, and asymmetric nominal. SAS calculates the similarity by means of the Procedure Distance. The equation, also provided in the Help section, is: v Ʃw • δ j s1 (x, y) = j x,y d • j=1 j x,y v Ʃw • δ j j x,y j=1 Where: v is the number of variables, or dimensionality wj is the weight of the jth variable where specified in the Procedure Distance, where wj = 0 when xj or yj is missing. s1(x,y) is the similarity between records x and y. For nominal, ordinal, interval or j ratio variables, we assume δ x,y = 1; For asymmetric nominal variables: j δ x,y = 1, if xj or yj are present; j δ x,y = 0, if xj and yj are missing. For nominal or asymmetric nominal variables: j dx,y = 1, if xj = yj j dx,y = 0, if xj ≠ yj For ordinal (where the corresponding rankings are used), interval or ratio variables: j dx,y = 1 - | xj - yj | The equation of the Gower general dissimilarity coefficient is given 11 as: d1 (x, y) = 1 ̶ s1 (x, y) 2.1.2 Clustering Algorithms According to Barroso and Artes (2003), most of the algorithms used for clustering purposes can be grouped into two main families of methods: hierarchical and non-hierarchical, also known as partition methods 2.1.2.1 Hierarchical methods In the application of hierarchical methods, clusters are formed based on a resemblance matrix (distances matrix). The first step in the algorithm is to identify the closest pair of observations. This pair is then grouped together to form the first cluster, and is regarded as a single record. The next step involves recalculating the distance matrix, given the change arising from the union of the pair of records in the preceding step. After recalculating, the closest pair is again identified and grouped, and so on successively until all the records have been grouped. What distinguishes between different hierarchic methods is the rule for redefining the resemblance matrix at each grouping of pairs. Some illus- trative methods follow: Single linkage method: The distance to be considered is the smallest distance between one element of groups G1 and G2, that is (equation 2). Complete linkage method: The distance to consider is the greatest distance between one element of groups G1 and G2, that is (equation 3). Ward method: Barroso and Artes (2003) report that, at each step, the Ward method attempts to link exposures to make the clusters formed as homogeneous as possible, using the total sum of squares partition from a variance analysis as a homogeneity metric. The Equation is (Equation 4): where SQT(X1) denotes the total sum of squares of the first information vector variable, SQE(X1) is the sum of squares between the groups (measuring the level of heterogeneity between groups) and SQD(X1) is the sum of squares within the groups (measuring the level of internal group homogeneity), Gj is the set of elements in group j, nj is the number of elements in group X X j, 1 is the measure of variable 1 and d (G1,G2)=min{d(i,k)}, where iЄG1e kЄG2 (2) d (G1,G2)=max{d(i,k)}, where iЄG1e kЄG2 (3) 12 SQT( X 1 ) = SQ E( X 1 )+SQ D( X 1 ) or k ƩƩ ( X j=1 1ЄGj k i1 -X 1 ) 2 = Ʃ n (X j=1 j X is the measure of the variable X1 in group j; therefore, the best initial partition for X1 is the one that minimizes SQD(X1) and, consequently, maximizes SQE(X1). The method calculates the sum of squares for every variable in the information vector and then calculates the partition’s sum of squares. The Formula is: j1 P SQDP= Ʃ SQD(i) i=1 where P is the total number of variables in the information vector The first step in the method’s algorithm involves forming (n-1) groups, with n as the total number of observations, and calculating SQDP for every possible cluster, choosing the one with the lowest SQDP. The next steps are forming (n-2), (n-3) to (n-k) groups, always taking the lowest SQPD as a selection criterion. Barroso and Artes (2003) note that, out of all hierarchical methods, Ward is the most attractive because it is based on a metric with strong statistical appeal and because it generates groups with high internal homogeneity. k j1 -X 1 ) 2 + ƩƩ ( X j=1 1ЄGj i1 -X j1 ) 2 , (4) 2.1.2.2 Non-hierarchical methods: k-means Partition methods attempt to find the partition whose groups show high internal homogeneity and are different from one another. The criterion used by the k-means method is based on the partition of the total sum of squares of a variance analysis, like the Ward method (Barroso and Artes, 2003). Mingoti (2005) summarizes that, in the k-means method, each element in the sample is allocated to the cluster whose centroid (sample means vector) is the closest to the observed values vector for the respective observation. The grouping algorithm begins with the a-priori definition of the desired number of groups and, based on this number, the same number of initial centroids, also known as seeds, is selected. The initial clustering is done based on the smallest distance between each exposure and the selected seeds, and the group’s centroid is recalculated after the union. The next step requires calculating the distance between each new record and the new centroids, attempting to determine whether the observations already allocated to one group are closer to other groups if affirmative, the observations are reallocated. 13 2.1.3 Comparison of Methods The main benefit of the k-means method is that it checks, with every step, whether the observations are allocated in the best way possible and allows reallocation when they are not. Under the Ward method, once two observations are grouped together, they will remain in the same group until the end of the procedure; in other words, the method does not take account of the fact that the inclusion of new observations in the groups may cause a certain observation to be closer to a neighboring cluster. Hierarchical methods, in turn, do not require aprioristic knowledge of the number of groups to be formed and, as a consequence, do not require setting the initial seeds. 2.1.4 Selection of the Number of Groups One of the main challenges in cluster analysis is defining the optimal number of groups to form. Several statistics exist to assist in this selection, such as the R 2 coefficient and the Pseudo-F statistic. According to Mingoti (2005), the Pseudo-F statistics runs an F test on variance analysis at every step to compare the mean vectors of the groups formed in the respective step. R2, in its turn, means the percentage of the total variability of the data that the clusters formed explain, as variability drops with the reduction of the dimension of the data. A higher Pseudo-F is desired, associated with a lower p-value for the test and, consequently, rejection of equality of the most significant mean vectors. For R 2 we would like values closer to one, so that little variability is lost by grouping observations. 3. Methodology For this application, we selected 20,000 overdraft limit contracts that were not in default, that is, less than 90 days late in June 2011. These contracts are part of the retail portfolio of a domestic financial institution. The exposures were tracked for the 12 following months, that is, until June 2012, and checked for lateness in excess of 90 days. Where such lateness occurred, the exposures were marked as in default. Application of the methodology used the software packages SAS 9.3 and SAS Enterprise Miner Workstation 7.1. 3.1 Variables Selection To obtain mutually heterogeneous groups in terms of risk of default, the variables that are most predictive of credit risk must be selected from those available. To this end, we used the SAS Enterprise Miner Workstation 7.1 software, Decision Tree node, to determine the most predictive variables for default based on the–log(p-value) statistic; the greater the statistic’s value, the lower its p-value and, therefore, the most predictive the variable. It is worth noting that the equation’s p-value is the p-value of the Chi-Squared test. 3.2 Variables Categorization Categorizing the selected variables is very important, as this will enable characterizing the groups formed 14 and subsequently allocating new exposures to those groups. Shiraiwa and Siqueira (2012) report that two different processes can be called variables categorization: the discretization of quantitative variables and the consolidation of qualitative variables. Discretization of quantitative variables consists of creating intervals so that each interval corresponds to one category. Qualitative variables are already presented as categories, although these categories may not meet the needs of the modeling process. For example, a variable may display too many categories, or some categories may display too small a number of observations. For the purposes of this study, the variables were categorized using the Chi-squared Automatic Interaction Detector, method, or CHAID, based on the SAS Enterprise Miner Workstation 7.1 statistical software. According to Kass apud Shiraiwa and Siqueira (2012, p.6) the method performs the following steps: 1. For each explanatory variable, draw a contingency table with the dimensions (c x d), where c is the number of categories for the independent variable and d is the number of categories for the dependent variable. 2. Find the pair of categories for the independent variable whose sub-table (2 x d) is significantly different. If the significance does not achieve the critical value, normally 5%, merge the two categories, regarding the merger as a single category, and repeat the step. 3. For each category made up of three or more original categories, find the most significant binary division. If significance is above a critical value, perform the division and return to step 2. 4. Calculate the significance of each consolidated independent variable and isolate the most significant. If the significance exceeds a critical value, sub-divide the data according to the (consolidated) categories of the selected independent variable 5. For each data partition that has not yet been analyzed, return to step 1. 3.3 Application of Cluster Analysis This study applied the following: • Gower’s General Dissimilarity Coefficient as a resemblance metric; • Ward’s hierarchical method as clustering method; • Pseudo-F and R2 statistics to define the ideal number of groups to form. 4. Results 4.1 Selected Variables Variables were selected by means of the application of a Decision Tree using the CHAID method, through the Decision Tree node of SAS Enterprise Miner Workstation 7.1. Every variable in table 1 was tested, but we decided to use a single variable for each of the dimensions required by the regulator, in order to ensure forming homogeneous groups based on the principle of parsimony. The results were as follows in Table 1. 15 ■ Table 1: Selected variables and value of the–log(p) statistic. Variable Description Dimension - log(p) perc_vr_utilizado Amount used / Contracted amount Operation 812,97 escore Behavior Score Obligor 739,41 atraso Days late Lateness 567,06 perc_crot_excesso Excess over limit / Contracted amount Operation 538,3 vr_excesso Excess amount t over limit Operation 527,17 qt_md_excesos Number of excesses over limit / Age of the checking account Operation 458,58 qtd_excessos Number of excesses over limit Operation 444,67 vr_utilizado Amount used Operation 308,08 tx_juros Interest rate Operation 78,1 vr_contratado Contracted amount Operation 42,24 Therefore, based on the–log(p) statistic, the following variables were selected: • Obligor Dimension: • Behavior Score (escore) • Lateness Dimension: • Number of days late on the month of reference (atraso); • Operation Dimension: • Percentage Amount Used (perc.vr.utilizado); The selected variables were also categorized using the CHAID method, Decision Tree node of the SAS Enterprise Miner Workstation 7.1 statistical package. The days late variable was categorized in a specialist manner, with a 30-day interval for each class. It is important to point out that the classes were arranged according to their percentage of default, so that class #1 corresponds to the one with the least percentage of defaults and so forth. The results were as follows (Table 2) 4.2 Categorization of the Selected Variables ■ Table 2: Categories for the selected variables. Behavior Score Days Late Pct Amount Used Class Intervals Class Intervals Class Intervals 1 >= 89,5 1 0 1 < 6,876 2 [70,5;89,5] 2 [1;30] 2 [6,876;21,08] 3 [55,5;70,5] 3 [31;60] 3 [21,08;54,665] 4 [43,5;55,5] 4 [61;89] 4 [54,665;72,013] 5 [14,5;43,5] 5 [72,013;100,002] 6 [2,5;14,5] 6 > = 100,002 7 < 2,5 16 4.3 Number of groups formed The results indicate the formation of seven groups, where the Pseudo-F value was 319,784.07 and R2 was 99.556%, as the Table 3 next shows ■ Table 3: Decision criterion for the optimal number of groups. ■ Table 4: Number of contracts per group. Group Number of Contracts Percentage of Total 1 5.756 28,78% 2 6.342 31,71% 3 2.272 11,36% 4 2.441 12,21% Number of Groups 5 1.568 7,84% R2 Pseudo F 6 471 2,36% 15 0,99481 294.487,46 7 1.150 5,75% 14 0,99313 262.780,86 13 0,99402 276.831,93 12 0,99098 243.921,66 11 0,98941 233.443,04 10 0,98669 211.781,58 9 0,99217 253.418,31 8 0,96326 104.846,90 7 0,99556 319.784,07 6 0,97514 130.693,86 5 0,93963 77.809,93 4 0,86544 42.869,59 3 0,78477 36.457,50 2 0,6565 38.220,22 1 0 . 4.5 Default rate by group The technique used to segregate the retail exposures portfolio into homogeneous risk groups is expected to form groups that are heterogeneous one in relation to another as concerns Probability of Default (PD). Since we have not yet provided a PD model to ascertain this difference, we calculated the Percentage Default (number of contracts in default divided by total contracts) for each group formed and observed the following: ■ Table 5: Percentage Default by Groups. 4.4 Number of contracts per group As for the distribution of contracts across the groups formed, Table 4 shows that the were well distributed, an no single groups shows too great a concentration. Group Percentage in Default 2 0,32% 7 1,57% 1 2,57% 3 6,07% 5 12,12% 4 22,33% 6 63,69% 17 As seen, the groups formed are significantly different in terms of Percentage of Default, supporting the hypothesis that the groups thus formed are different in terms of risk of default. The smallest difference (1%) was found between groups 1 and 7. 4.6 Specification of the Groups Formed To specify the groups formed for later classification of new exposures into one of the um dos homogeneous groups, we must find the frequency of the class vector of the variables that constitute each group. This class vector has been termed characteristics vector and is made up of the concatenation of the Behavior Score, Days Late and Percentage Amount Used variables, respectively. Cluster Analysis ensures that exposures with the same characteristics vector will always be in the same group, as they are the first to be grouped together by the clustering algorithm, given that the distance between them is zero. The results were: ■ Table 6: Characteristics vectors by group. Group Characteristics Vectors 1 211, 311, 411, 511, 611, 711 2 111 3 212, 213, 214, 215 4 315, 415, 515, 615, 715 5 312, 313, 314, 412, 413, 414, 512, 513, 514, 612, 613, 614, 712, 713, 714 6 126, 226, 241, 326, 436, 525, 526, 536, 546, 626, 636, 646, 726, 736, 741, 746 7 112, 113, 114, 115 4.7 Interpretation of the groups formed Table 6 contains the characteristics vectors and allow interpreting the groups formed. Generally speaking, the lateness variable is a strong characteristic for the identification of risk of default in a credit portfolio, and so is the percentage amount used variable, given that as the credit limit is used up, the risk increases of a customer not honoring repayment and becoming late. However, it is interesting to point out that the obligor risk characteristic, represented by the Behavior Score, was also significant for the definition of homogeneous risk groups. This can be seen in group 1, which is formed by exposures that display class 1 for the lateness and operation dimensions, that is, contracts without late payments and whose usage is up to 6.876%, and with BS scores below 89.5, on a 0-100 scale. Group 2, in its turn, is made up only of exposures that are class 1 for all dimensions, representing the lowest risk group, comprehending contracts without late payments, whose usage is of up to 6.876%, and whose Behavior Score is over 89.5 and under 100. Table 5 shows that this is the group with the smallest percentage of defaults, at 0.32%. Group 3 is made up of exposures with Behavior Scores between 70.5 and 89.5 (class 2) on the obligor dimension, zero days late (class 1), and with operation dimension classes ranging between 2 and 5, that is 6.876% to 100.002% credit limit usage. Note that the obligor and lateness dimension classes, respectively 2 and 1, are low-risk classes, which re- 18 Authors flects on the group’s default percentage of 6.07% In group 4, the obligor dimension classes range from 3 to 7 (Behavior Score below 70.5), while the lateness dimension class is 1 (zero days late) and the operation dimension’s is 5 (72.013% to 100.002% usage). The group’s classes for the obligor and operation dimensions are riskier, impacting the group’s percentage default of 22.33% (Table 5). In group 5, the obligor dimension also ranges from 3 to 7 (BS under 70.5), while the lateness dimension class is 1 (zero days late) and the operation dimension’s classes go from 2 to 4 (6.876% to 72.013% limit usage). As the data show, the class variations are rather similar to group 4, but group 5’s obligor dimension classes are less risky and this leads to a percentage of default of 12.12%, below group 4’s. In group 6, the obligor dimension’s classes vary between 1 and 6 (89.5 to 14.5), while the lateness dimension’s classes range from 2 to 4 (1-89 days late) and the operation dimension’s classes are between 2 and 4 (6.876% to 72.013% limit usage). This group showed the most variation in terms of the dimension’s classes, which may have reflected on its 63.69% defaults percentage. In group 7, exposures were class 1 (the best class) for the obligor and lateness dimensions (BS 89.5 or higher and zero days late) and the operation dimension’s classes varied between 2 and 5 (6.876% to 100.002% limit usage), indicating that the obligor and lateness dimensions contribute the most to the allocation of exposures to the group. As Table 5 shows, this is the group with the second lowest percentage of defaults, at 1.57%. 5. Final Comments The clusters formed showed good distribution in terms of the number of contracts, were consistent based on characteristics vectors analysis, and quite distinctive in terms of percentage of default, which is one of the most important traits to support the hypothesis of using cluster analysis to categorize retail exposures according to homogeneous risk groups. Therefore, the study’s main objective was reached and it has been proven that Cluster Analysis is a viable option to segregate a retail exposures portfolio into homogeneous risk groups. One topic future studies may address involves developing a methodology to allocate exposures with observation vectors not found in the development sample, or to classes that may be created over time in one of the homogeneous groups formed. A conservative way of dealing with this problem is to allocate such exposures to the worst group, that is, the one with the highest percentage of defaults. Fabio Augusto Scalet Medina Bachelor of Statistics and Specialist in Applied Statistics, currently Senior Executive Assistant at the Credit Risk and Collection Modeling Area of a domestic retail bank. Contact: [email protected] Authors 19 Edson Luiz de Carvalho Barbosa Master of Production engineering with emphasis on Operational Research from COPPE/UFRJ and Bachelor of Statistics from UERJ. Has an MBA in Risk Management from FIPECAFI. Currently National Credit Risk and Collection Modeling manager in a domestic retail bank. Bibliography Contact: [email protected] ASSAF NETO, A. Mercado Financeiro. 10ª ed. São Paulo: Atlas, 2011. 339 p. BARROSO, L. P.; ARTES, R. 10º Simpósio de Estatística Aplicada à Experimentação Agronômica (SEAGRO), 2003, Lavras. Análise Multivariada: Minicurso. Lavras: UFLA, 2003.151 p. BIS - Bank for International Settlements. Disponível em: <http://www.bis.org/bcbs/history.htm>. Acessado em 30/01/2013. CARNEIRO, F. F. L.; VIVAN, G. F. A.; KRAUSE, K. Novo Acordo da Basileia: estudo de caso para o contexto brasileiro. Resenha BM& F, São Paulo, n.163, 2005. Disponível em < http://www4.bcb.gov.br/pre/inscricaoContaB/trabalhos/O%20Novo%20Acordo%20de%20Basil%C3%A9ia_um%20estudo%20de%20caso%20para%20 o%20contexto%20brasileiro.pdf>. Acessado em 05/02/2013. Circular BACEN nº. 3.360 de 12 de setembro de 2007. Disponível em <http://www.bcb.gov.br/pre/normativos/busca/normativo.asp?tipo=circ&ano=2007&numero=3360>. Acessado em 05/02/2013. Circular BACEN nº. 3.581 de 8 de março de 2012. Disponível em <http://www.bcb.gov.br/pre/normativos/busca/ normativo.asp?tipo=circ&ano=2012&numero=3581>. Acessado em 05/02/2013. JOHNSON, R.A.; WICHERN, D.W. Applied Multivariate Statistical Analysis. 4th ed. Upper Saddle River: Ed. Prentice Hall, 1998. 816 p. JORION, P. Value at Risk: a nova fonte de referência para a gestão do risco financeiro. São Paulo: BM&FBOVESPA: Bolsa de Valores, Mercadorias e Futuros, 2010. 487 p. LANDIM, P.M.B. Análise estatística de dados geológicos multivariados. Lab. Geomatemática, DGA, IGCE, UNESP/Rio Claro, Texto Didático 03, 128 p. 2000. Disponível em <http://www.rc.unesp.br/igce/aplicada/DIDATICOS/LANDIM/multivariados.pdf>. Acessado em 30/01/2013. MENDONÇA, A. R. R. de. O Acordo da Basiléia de 2004: uma revisão em direção às práticas de mercado. Economia Política Internacional: Análise Estratégica, Campinas, Instituto de Economia, Unicamp, n. 2, jul./set. 2004. Disponível em: <http://www.eco.unicamp.br/ceri/boletim2.htm>. Acessado em 02/02/2013. MINGOTI, S.A. Análise de dados através de métodos de estatística multivariada: Uma abordagem aplicada. Belo Horizonte: Ed. UFMG, 2005. 297 p. RAPELI, C. B.; BOTEGA, N. J. Clinical profiles of serious suicide attempters consecutively admitted to a university-based hospital: a cluster analysis study. Revista Brasileira de Psiquiatria, nº 27, 285-289, 2005. Disponível em <http://www.scielo.br/scielo.php?pid=S1516-44462005000400006&script=sci_arttext>. Acessado em 02/02/2013. Bibliography 20 Resolução CMN nº. 3.490, de 29 de agosto de 2007. Disponível em <http://www.bcb.gov.br/pre/normativos/busca/ normativo.asp?tipo=res&ano=2007&numero=3490> Acessado em 02/02/2013. SANVICENTE, A. Z.; MINARDI, A. M. A. F. Migração de risco de crédito de empresas brasileiras: uma aplicação de análise de clusters na área de crédito. Instituto Brasileiro de Mercado de Capitais, Working Paper, março de 1999. Disponível em <http://www.risktech.com.br/PDFs/migra%C3%A7%C3%A3o.pdf>. Acessado em 02/02/2013. SAS Help. Coeficiente de Similaridade e Dissimilaridade geral de Gower. Disponível em <http://support. sas.com/documentation/cdl/en/statug/63962/HTML/default/viewer.htm#statug_distance_sect008.htm>. Acessado em 02/02/2013. SAS Help. Fórmulas do Coeficiente de Similaridade e Dissimilaridade geral de Gower. Disponível em <http:// support.sas.com/documentation/cdl/en/statug/63962/HTML/default/viewer.htm#statug_distance_sect016. htm>. Acessado em 02/02/2013. SHIRAIWA, C. S.; SIQUEIRA, J.O. Categorização de Variáveis Em Modelos de Risco De Crédito, Artigo referente à conclusão de MBA em Gestão de Riscos da Fundação Instituto de Pesquisas Contábeis, Atuariais e Financeiras – FIPECAFI, 2012, 13p. 21 Survival Analysis in Credit Models: a View of Covariates Treatment Melissa Sousa Tiago Lima Wanderson Rocha 22 Abstract The study proposes using survival analysis to adjust a credit model, with particular focus on the treatment of its predictive variables, as well as on the creation of homogeneous clusters. Unlike other studies that use survival analysis in connection with credit data, this study’s highlight is a concern with the relationship between output and predictive variables. We refer to output variables in the plural because we understand that two main aspects exist in the credit arena: the default event and time to default; the former captures the core concern of credit models – default -, while the latter is concerned with when an individual will default. We use the nonparametric Cox model to adjust the model so that time to default and default rate are aligned. The logrank test was used at two stages in the study: treatment of the predictive variables, to ensure categories with distinct survival curves, and the creation of de homogeneous clusters that, in addition to different default rates over time, also show different survival curves. The study has been divided into four sections, with an introduction of the problem at hand, a view of the variable-treatment process, the adjustment of the model via Cox and, finally, the method for creating homogeneous clusters. KEYWORDS: Survival analysis, Cox Model, Logrank Test, Variables Treatment. 1. Introduction Several studies exist on the use of the survival analysis technique in credit applications. The output variables that regulatory authorities defend are mostly concerned with the observation of an event over a certain period of time, which makes for binary outputs. However, if an institution knows, in addition to the rate at which an event will occur in the coming months, the form of the distribution over this period, such an institution will be able to improve its strategy and increase financial returns. One of the solutions frequently suggested in such cases is survival analysis. In the majority of the discussions about the topic, the main fo- 23 cus lies on the selected technique and the evaluation of its performance. Variables treatment ends up in a secondary role and, often, the same treatments used for logit regression models are used, based on a binary output variable. In practical application however, the treatment of variables requires plenty of attention, particularly where one wants a single model to provide the regulatory probability of default and the timing of its distribution. It is important to make sure that the independent variable identifies the behavior of the two elements of the output. This study proposes a joint treatment of three variables: the relevant predictive variable, the regulatory default rate, and the time to default. Treatment is important because there may be variables with distinct behaviors for each of the outputs. For example, a certain age group may show a higher rate of default, but with a longer time-to-default than another group with a lower default rate. ■ Table 1: Variable categorization example Age Time to default (days) Default rate (12 months) Up to 30 250 5.0% 30 - 40 230 4.0% 40 - 50 200 3.0% 50 - 60 180 2.0% 60 - 70 140 1.0% Over 70 110 0.5% If the categorization only takes account of default at a certain point in time, the variable will not make the best contribution possible to the model when it comes to estimating the event over different spans of time. 2. Modeling Process The modeling process we propose in this study assumes that the population sampling and segmentation stage has been completed. In a three-stage process, we begin with the analysis of predictive variables and proceed to the construction of homogeneous clusters, as proposed by the regulators. The stages are: • Variables treatment; • Model adjustment and performance analysis; • Construction of homogeneous clusters. 2.1 Variables Treatment The first stage in the process attempts to fully understand the predictive variables in connection with the default event and time to default. To adjust the credit risk model, we must make sure that its predictive variables are consistent. This evaluation is done based on the predictive variable and its relationship with defaults. All covariates are evaluated, regardless of type (numeric or categorical); the difference lies in the kind of summarization used: mean and median are considered for numeric variables, while categorical ones rely on frequency. The treatment of the covariates for the purposes of risk model adjustment abides by a pre-selection analysis flow. As each 24 covariate meets a criterion, it is brought to the following phase. 2.1.1 Variability and Stability Assessment The first pre-selection criterion measures whether the covariates have many missing values or concentrate on a certain value, in addition to checking the temporal behavior of the relationship between each covariate and defaulting. 2.1.2 Categorization Analysis The bivariate analysis method is widespread among Analytics teams; generally speaking, it involves categorizing the continuous covariate so that each category makes sense when it comes to evaluating defaults. Adjusting a survival model does not greatly depart from this, but the analysis of the covariate’s meaning must be performed for both default and time to default. We propose to divide the ordinal continuous covariates into deciles, whe- re each decile is regarded as a stratum for the survival curve’s adjustment. The goal, here, is to check whether each of the variable’s strata (categories) produces distinct survival curves where each curve is monotone in meaning relative to default and time to default. We used the Logrank test to compare the consecutive categories; categories were reclustered where statistically equal. The test was repeated until every category showed a statistically different survival curve. After the pre-definition of categories, we simultaneously compared the default rates by category and the median time to default. For the model to correctly provide the rate of default for each observation period, the two indicators had to be aligned and with opposite trends. Graphic 1 illustrates the analysis. Graphic 1 (A) shows the behavior of the default rate and median time to default observed in the pre-catego- ■ Graphic 1: Default rate and median time to default by category. 25 rization. As seen, we were unable to obtain monotone effect with 10 categories. We achieved the effect by reclustering categories [2 and 3], [5 and 6] and [8 and 9], as Graphic 1 (B) shows. After defining the categories, a Kaplan-Meier model and a Logrank test were again run using the new categories. 2.1.3 Category Stability Analysis Once the categories were defined for each covariate, we investigated whether they were stable over time. This stage concerns itself with default and each category’s share. The main point at this stage is to check for variables with a volume of individuals per category such that the default rate is not time-dependent, that is, stationary. Visual analysis is recommended for this stage, and Graphic 2 illustrates the ■ Graphic 2: Default kind of graph proposed for visualization, considering a covariate with 4 final categories. 2.2 Model Adjustment and Performance Analysis Different survival analysis methodologies exist. This study focuses on nonparametric adjustment, using the Cox model. Performance analysis was an innovation on the part of the study, as this chapter will show. 2.2.1 The Cox Nonparametric Model The Cox regression model allows analyzing data from lifetime studies where the output is the time until the occurrence of a relevant event, with adjustment via covariates. Generically, considering p covariates, so that x is a vector with the rate and volume stability by category. 26 components x=(x1, x2,...,xp), the general expression of the Cox Model is: se the ratio between the failure rates of two different individuals is constant over time. The ratio of the failure rate functions for individuals i and j is given by (Equation 2) and is not time-dependent. λ(t ) =λ 0 (t )g (x´β) where g is a non-negative function that must be specified so that g(0)=1. This model is made up of the product of two components, of which one is nonparametric and the other is parametric. The nonparametric component λ0( t ) is not specified and is a non-negative function of time. It is usually referred to as baseline, sinceλ( t )=λ0 when x=0. The parametric component is often used in the following multiplicative form (Equation 1): where β is the parameters vector associated with the covariates. This ensures that λ( t ) is always non-negative. The constant β0 is not shown in the Equation above because of the presence of the model’s nonparametric component, which absorbs the constant term. The Cox model is also referred to as the proportional hazards model becau- 2.2.2 Cox Model Adjustment The Cox regression model is characterized by the β´s coefficients, which measure the effects of the covariates on the failure rate function. An estimation method is needed to allow inferences about the model’s parameters. The most frequently used method is the Cox partial maximum likelihood model, which consists of conditioning the construction of the likelihood function on awareness of the past history of failures and censures. In a sample of n individuals with k = ≤ distinct failures at times t1 < t2 ... < tn , the conditional probability of the i th observation failing at time ti is given by (Equation 3): where R ( ti ) is the set of the indices of observations at risk at time ti . g(x´β) = exp{x´β}= exp{β1 x1+...+βp xp } λ i (t ) λ j (t ) = λ 0 (t )exp{x i ´β} λ 0 (t )exp{x´ j β} = exp{x´i β –x´ j β}, (1) (2) P [ individual failure at ti | One failure at ti and history until ti ] = λ 0 (t )exp{x i´ β} = exp{x i´ β} Ʃj Є R (t ) λ 0 (t )exp{x j´ β} Ʃj Є R (t ) exp{x j´ β} i i (3) 27 The likelihood function to be used for the purposes of inferring the model’s parameters is therefore given by the product of all of the terms presented, associated with their distinct failure times, where δi is the failure indicator (Equation 4). k L (β) = Π i=1 Ʃ λ 0 (t )exp{x i´ β} λ (t )exp{x j´ β} j Є R (t ) 0 i 2.2.3 Model Performance The literature discusses several methods to evaluate the fit and quality of the Cox Model. These methods are essentially based on the residuals of Cox-Snell, Schoenfeld, Martingal and Deviance. The firmer two evaluate the fit of the assumption of proportional hazards and the latter two evaluate atypical points and the covariates’ functional form. Proportional hazards assumption is often not applicable to the variables credit models study. Given this, validations are required in terms of the model’s estimated values. Evaluation of the performance of a model with a given portfolio, with the purpose of direct application to credit management, is given by the metrics practitioners generally use: Kolmogorov Smirnov (KS), Gini Coefficient, and AUROC (Area under Receiver Operating Characteristic). A new proposition that may be adopted based on the construction of survival models is measuring these statistics over time, that is, over different evaluation periods, such as 3, 6, 12 and 24 months, for example. One important aspect in the evaluation of the estimated model is the comparison of estimated and observed rates of default. The estimated rates must not underestimate the portfolio’s actual rates, which also applies to the different n = Π i=1 (Ʃ λ 0 (t )exp{x i´ β} j Є R (t i ) δi ) λ 0 (t )exp{x j´ β} (4) evaluation periods. Therefore, a graphic method to evaluate the observed and estimated default curves for different observation periods can be used to validate the final model. The estimated rate of default for the observation of the model over time us given by the following relationship according to the Cox model: estimed rate = 1 – S0(t), Where S0(t) is given by: S (t) = λ 0 (t)*e x´β Therefore, in addition to the estimated βj parameters, we must obtain the baseline λ0 (t) for every t . For example, the estimated rate of default for 365 days would be given by: estimed rate (365)= λ 0 (365)* ex´β Below, we provide an illustration of the estimated and observed rates over the previously defined times (Graphic 3): 28 ■ Graphic 3: Observed and expected rates. 2.3 Construction of Homogeneous Clusters According to Public Hearing Notice No. 37, published in 2011 by the Central Bank of Brazil, a homogeneous cluster may be defined as a “set of exposures or obligors with shared characteristics for the purposes of credit risk evaluation and quantification”. Homogeneous clusters act as facilitators for the management of the institution’s credit portfolio. The constitution of these clusters must take into consideration certain specific criteria: • A sufficient number of observations per cluster. The observed rate of default must be an efficient PD estimator (non-biased and with minimum variance); • The rate of default must be monotone rising or falling as a function of the direction of each cluster’s score; • The rates of default obtained for each cluster must be statistically different; • There can be no concentration of observations in a specific cluster. • Each cluster’s rates of default must show distinct levels across the analyzed vintages and must also be stable. The means practitioners use most frequently to build these clusters relies on breaking the development sample into percentiles so that the rates of default are ordered. These breaks are then reclustered. The most frequently used tests to measure differences between the clusters’ default rate levels are the Duncan, Tukey and Scheffé tests. To check for the rate’s stability across vintages, they use the Population Stability Index (PSI), so that distribution shifts can be measured. The Herfindahl test can be applied to measure the concentration of observations in the clusters formed. After the homogeneous clus- 29 ters are defined, each cluster’s empirical survival curve can be studied. These curves are compared using the Log-Rank test. Where the hypothesis of equality between two curves is not rejected (p-value over 5%), the breaks are reclustered and the tests are run anew. The Graphic 4 and Table 2 below illustrate the test and the clusters’ behavior over time relative to the survival curve. Graphic 4 illustrates the survival curves of five homogeneous clusters found. Table 2 shows the Logrank test to check whether all survival curves are distinct when paired. 3. Conclusion When applied to credit risk models, survival analysis usually performs well from the angle of explaining de- ■ Graphic 4: Survival curves by homogeneous cluster. ■ Table 2: Logrank test – illustrative. Cluster Cluster Chi-squared p-value 1 2 q12 <0.0001 1 3 q13 <0.0001 1 4 q14 <0.0001 1 5 q15 <0.0001 2 3 q23 <0.0001 2 4 q24 <0.0001 2 5 q25 <0.0001 3 4 q34 <0.0001 3 5 q35 <0.0001 4 5 q45 <0.0001 fault at a specific point in time – considering curve after 365 days, in this case. The methodology’s benefit compared to the usual logit regression is the ability to preview defaulting individuals over different horizons. The main point of improvement was the ability to add the traditional practitioner concept with the concept of time to the event, which was possible because all the analyses done on the predictive variables focused on the two aspects, lending stability and accuracy to the adjustment. Authors 30 Melissa Sousa Has a Bachelor’s degree in Statistics from Universidade Estadual de Campinas and an MBA in Finance from FIA/USP. Currently an Analytics consultant at Serasa Experian, her experience includes credit risk modeling, CRM, and Analytics solutions. She has worked in modeling areas devoted to several kinds of portfolios, focusing on financial performance, always in the banking industry. Sousa can be contacted by e-mail: [email protected]. Tiago Lima Has a Bachelor’s degree in statistics from Universidade Estadual de Campinas and a Master’s degree in statistics from Universidade de São Paulo. Currently an expert in econometric credit risk models development. His experience includes time-series projection models and multivariate analysis. Lima can be contacted by e-mail: [email protected]. Wanderson Rocha Has a Bachelor’s degree in Statistics from Universidade Federal do Paraná and is a post-graduate student of Analytical Intelligence at FIA/USP. He has been with the Serasa Experian Analytics area since 2012. His experience includes credit risk modeling and Basel II in particular. Previously, he worked in a major financial institution’s Corporate Credit Risk – Basel II area for 3 years. Wanderson can be contacted by e-mail: [email protected]. 31 Spatial Correlation of Corporate Defaults Guilherme B. Fernandes Rinaldo Artes 32 Abstract In recent decades, measuring credit risk has been a constant concern for financial institutions; credit scoring are important tools for this purpose. The national and local economic situation may affect a company’s or an individual’s credit risk. Several indices exist to measure a country’s economic risk, but the same is not true of business and individual risk. Local economic effects may provide information on the default rates of companies or people that operate/live in the respective areas; this can be measured using spatial statistical analysis techniques. In this paper, we use Moran’s I to evaluate the spatial correlation of default for companies in the State of São Paulo. We perform a time analysis and suggest a structure associated with the granularity level. The main results are evidence of the presence of a spatial correlation and the relationship between Moran’s I and the market rate of default. Keywords: Spatial Correlation, Rate of Default, Moran’s I, Credit Risk. 1. Introduction One of the objectives of Basel II (BCBS, 2006) is to encourage measurement of the credit-risk to which each financial institution is exposed and propose calculating the capital to be allocated in order to prevent against losses. The Vasicek (2002) model is regarded as the basis for the required capital formula. The model assumes that a company’s probability of default (PD) depends on its idiosyncratic characteristics and on a macro-economic factor common to all companies. Small and medium-sized businesses in particular are more heavily influenced by regional/local economic factors. The effect can be observed in small areas such as region, sub-re- gion, or even sector and sub-sector, according to Brazilian Post Office (“Correios”) postal code (“CEP”) digit classification (http://www.correios.com.br/ servicos/cep/cep_estrutura.cfm). The economic activity indices currently published in Brazil do not reach this level of information capillarity. Therefore, inclusion of these effects into a PD model is not possible. However, although such regional factors cannot always be observed, their effects can be identified by means of spatial analysis of default. In this paper, we use Moran’s I to study the evolution of the spatial dependence of the default of small and medium-sized companies between São Paulo state regions defined according 33 to CEP numbers, from December 2007 to August 2010. We considered three regionalization criteria. In addition, we correlated Moran’s I with SME default rates observed in the state of São Paulo during the relevant period. 2. Literature Review Probability of default may be conditioned on several risk factors, of which the location of an applicant relative to other obligors may be a potentially relevant factor, as Stine (2011) points out. In this article the author analyzes real-estate credit risk in the United States pre- and post-2008 crisis, based on county-level default rate data. Figure 1 was extracted from this paper and shows the change in default in 2005-2009 (the full paper analyzes the 1993-2010 period). Figure 1 uses warm colors to represent counties with high default rates and cool colors for those where the rates are low. Spatial default-rate correlation is clearly visible. Default rises intensively in the West Coast during period. In this case, a county’s rate increase appears to be related with the increase in neighboring counties. Cotterman (2001) analyzes the impact of local characteristics on consumer property credit risk in the United States. The neighborhood’s ethnic make-up and income are associated with default risk factors. Cotterman (2001) thus attempts to capture the causative effect of the spatial correlation based on neighborhood socio-demographic characteristics. ■ Figure 1 Evolution of default in the USA by county. (Source: Stine - 2011) 34 The paper’s conclusions include evidence that the effects of local factors (income and race) lose strength when the model is conditioned on other characteristics for the applicant, operation and macro-economic environment. Cotterman’s paper focuses specifically on property credit securitized by the Federal Housing Administration (FHA), the US Government agency for mortgage securitization. Cotterman (2001) incorporates the spatial correlation by means of postal code clusters. In the following years, several papers came out on the spatial correlation of credit risk, but always conditioned on the distance between applicants and creditors. Grunert and Weber (2009), Degryse and Ongena (2005), Carling and Lundberg (2005) and, later, Argawal and Hauswald (2007) use empirical evidence to approach the correlation between default rates and the applicant-creditor distance. Deng, Pavlov and Yang (2005) model the cause of mortgage foreclosure and incorporate spatial heterogeneity into their model. Some of the variables that influence foreclosure are often latent: culture, access to information, etc. According to Deng, Pavlov and Yang (2005), individuals with those latent characteristics agglomerate in neighborhoods, which enables modeling the cause of foreclosure. Deng, Pavlov e Yang (2005) address three causes of foreclosure: refinancing, property sale and default. Non-foreclosed contracts are regar- ded as censures. In this paper the model assumes that each individual has a choice: refinance, sell, or default. The correlation structure between co-variables and the response variable depends on the type of neighborhood, through which the authors introduce spatial information into the model. Deng, Pavlov and Yang (2005) propose a competitive risks model for the causes of mortgage foreclosures. The model’s parameters are variant, in an attempt to capture the effect of latent variables shared by individuals in a certain region. In a 2012 paper, Agarwal et al. identify the presence of spatial default correlation by analyzing mortgages in the US. Their study estimates the increase in the risk of default of residents of regions with higher foreclosure rates. However, controlling for other individual factors, the concentration of sub-prime mortgages does not increase the credit risk of a non-defaulting neighbor. Only higher risk operations such as hybrid ARMs (Fabozzi, 2006) or documents with incomplete documentation show this spatial correlation. In essence Argawal et al. (2012), Deng et al. (2005) and Cotterman (2001) propose the presence of latent unobserved factors resulting in a spatial correlation among individuals. Stine (2011), on the other hand, proposes a pragmatic analysis of the spatial correlation structure via the spatial correlation for areas known as Moran’s I (Moran, 1950). 35 3. Methodology Type 1) wik = 1/dik, where Moran’s I (I-Moran) is a correlation metric that incorporates the two dimensions involved in a spatial phenomenon: latitude and longitude. As defined in Moran (1950), the index is given by the Equation 1: I = ƩƩ N ƩƩ i i k w ik where yi is the yth observational unit’s default, ȳ is the mean default, N is the number of observations and wik is the weight related to the distance between units i and k. Note that wik lacks a rigid, single form; therefore different structures will result in different Moran’s I values. Moran’s I is widely used in studies where the observational units are areas. Some illustrative examples of w_ ik given this context include: dik is the distance between observations i and k. 2 Type 2) wik = 1/dik , { Type 3) wik= k 1, where i and k are neighboring regions 0,otherwise w (y i – ȳ) (y k – ȳ) ik Ʃ (y – ȳ) i i 2 , (1) 3. Results Stine (2011) shows that for individuals, revolving, installment and mortgage credit portfolios show a strong spatial correlation, with Moran’s I ranging between .25 and. 65, depending on the portfolio and the period. Graphic 1 shows this variation over time. It is worth pointing out that the spatial correlation depends on the macro-economic scenario. The 2000-2006 period shows increased spatial correlation, but ■ Graphic 1 Spatial correlation – Moran’s I (Source: Stine - 2011) 36 even after the 2008 crisis, although there is a drop in Moran’s I, spatial default-rate dependence persists among US counties. Stine (2011) formulated the weights schedule as being type three with neighborhoods of up to two layers, that is, adjoining counties or counties separated by up to one county have weight 1. In this paper, we evaluate Moran’s I using type-2 weight, that is, the inverse of the squared distance between regions. In this case, the region’s centroid was used to calculate distances. The first selected region level was the CEP region, represented by the code’s first two numerals. Figure 2 shows the State of São Paulo’s 19 regions. The first map shows the nine coastal and interior regions, the second shows the six greater São Paulo regions, and, finally, the city of São Paulo’s fur regions. Additional details on the regions denoted by the first two CEP numerals can be found in MundoGeo (2013) and NeoGeographiká (2013). However, analysis of Moran’s I for these 19 regions in the relevant period showed low spatial correlation results. Graphic 2 shows that Moran’s I does not exceed .1 in this case. Analyzing sub-regions, that is, areas determined by the first three CEP numerals, spatial correlation increases and ranges between .15 and .21. Finally, when sector (first four CEP numerals) is ■ Figure 2 Illustration of nineteen regions of the state of São Paulo. ■ Graphic 2 Moran’s I by CEP granularity level. 37 taken into consideration for Moran’s I calculation purposes, the correlation rises to around .55. Greater granularity would require calculating the distances among 20 thousand areas and be subject to computing limitations. As in Stine (2011), Moran’s I varies over time. In our study, variation over time is smaller due to the shorter period involved: 3 years, as opposed to 19. Still, one can discern a rising trend in spatial correlation (sector and sub-region levels) over time. The correlation between market default rate (BACEN Website, 2013) and Moran’s I is of around .65. In other words, there are indications that increased spatial correlation occurs together with increased market defaulting. As for SMEs, the presence of spatial correlation is intuitively consistent. Small and medium-size companies (SMEs) in particular show greater spatial correlation with nearby business firms. Firms of this size exist to serve their region’s population and larger companies. This may result in spatial dependence, as a cooling economy and the resulting increase in defaults cause the credit risk of a given company in the region to rise. 4. Closing remarks The spatial correlation among small and medium-sized enterprises may be due to a failure to observe local economic activity factors. The correlation level depends on the definition of the analyzed regions and this paper found a weak spatial correlation as measured by Moran’s I for the 19 São Paulo state regions (the first two CEP numerals). However, when the area definition is based on the state’s sectors (first four CEP numerals), Moran’s I indicates the presence of significant spatial correlation. Should this information be disregarded while developing a credit risk model – the so-called biased due to the endogeneity created by omitted variables. Furthermore, it is worth noting that the correlation between the rate of default and Moran’s I is moderate, with a Pearson coefficient of .65. A cause-and-effect relationship does not necessarily exist, but in periods of high defaulting, credit scoring models that fail to incorporate the spatial correlation present in the data will be more liable to lose performance. One topic for future research is how to include information from the spatial correlation into credit scoring models. One direct means would be through the variances and co-variances matrix, but in this case the issue of computing limitation must be taken into consideration. Another front, approached by Fernandes (2012), would be to estimate spatial risk through techniques such as Kriging. Finally, the time correlation found between spatial dependence and market rate of default suggests a future detailed study to evaluate the vulnerability of credit portfolios. Authors 38 Guilherme B. Fernandes Guilherme Fernandes has a Bachelor’s degree in Statistics from Universidade Federal de São Carlos and a Master’s degree in Economics from Insper-SP, and is a Doctor of Statistics candidate at ICMC-USP/DEs-UFSCar. He is currently responsible for Serasa Experian’s Analytics Innovation area, where he has been since 2010. His experience focuses on credit and fraud risk modeling. He was previously employed for four years in the corporate credit risk area of a major Brazilian bank. Fernandes may be contacted by email at [email protected]. Prof. Dr. Rinaldo Artes Prof. Dr. Rinaldo Artaes is a Bachelor, Master and Doctor of Statistics from Universidade de São Paulo. He is currently a full-time professor at Insper Instituto de Ensino e Pesquisa. His main areas of research include References multivariate analysis, circular data and estimation equations. BAGARWAL, S. E HAUSWALD, R.B.H. (2007): Distance and information asymmetries in lending decisions. Proceedings, Federal Reserve Bank of Chicago, issue May, pp. 183-204. AGARWAL, S.; AMBROSE, B.W; CHOMSISENGPHET, S. e SANDERS, A.B. (2012): Thy Neighbor’s Mortgage: Does Living in a Subprime Neighborhood Affect One’s Probability of Default?, Journal of Real Estate Economics, American Real Estate and Urban Economics Association, vol. 40, No. 1, pp. 1-22. Banco Central do Brasil: Consulta ao site https://www3.bcb.gov.br/sgspub/localizarseries/ às 13:50 do dia 28/Abril/2013. Série utilizada: Operações de crédito aos setores público e privado - recursos livres - inadimplência – PJ. Basel Committee on Canking Supervision (2006): International Convergence of Capital Measurement and Capital Standards – A Revised Framework. Bank of International Settlements. CARLING e LUNDBERG (2005): Asymmetric information and distance: an empirical assessment of geographical cretit rationing, Jounal of Economics and Business. Vol. 57, pp. 39-59. COTTERMAN, R.F. (2001): Neighborhood effects in mortgage default risk. Relatório encomendado por U.S. Department of Housing and Urban Development. Download feito de http://www.huduser.org/portal/publications/hsgfin/defaultrisk.html na data de 20/03/2012. DEGRYSE, H. e ONGENA, S. (2005): Distance, Lending Relationships, and Competition, Journal of Finance, American Finance Association, Vol. 60, No. 1, pp. 231-266. DENG, Y.; PAVLOV, A.D. e YANG, L. (2005): Spatial Heterogeneity in Mortgage Terminations by Refinance, Sale and Default, Journal of Real Estate Economics, American Real Estate and Urban Economics Association, vol. 33, No. 4, pp. 739-764. FERNANDES, G. (2012): Mensuração do risco de crédito espacial e sua incorporação nos modelos de credit scoring. Dissertação de Mestrado, Insper,São Paulo, SP. GRUNERT, J. e WEBER, M. (2009): Recovery Rates of Commercial Lending: Empirical Evidence for German Companies. Journal of Banking and Finance, Vol. 33, pp. 505-513. MORAN, P.A.P. (1950): Notes on Continuous Stochastic Phenomena. Biometrika. Vol 37, No. 1, pp. 17–23. MundoGeo (2013): Acesso feito à página http://mundogeo.com/blog/1998/08/02/geonegocios-8/ em 30 de abril de 2013. Neogeographiká (2013): Acesso feito à página http://neogeographika.blogspot.com.br/2012/08/ken-mapas-e-plantas. html em 30 de abril de 2013. STINE, R. (2011): Spatial temporal models for retail credit. Credit Scoring and Credit Control Conference 2011, Edimburgo, Reino Unido. 39 Infrastructure Credit and Financing in Brazil Frederico A. Turolla Márcio F. Gabrielli Igor J. C. Gondim 40 Introduction A window of opportunity has opened for a leap ahead in infrastructure investment after the developments for the Brazilian economy in the 1990s and 2000s. In this period, and in the 1990s in particular, the country underwent a ranging review of key elements in its institutional environment. The sweeping reforms addressed areas of broad impact, such as prices stabilization, the introduction of more modern competition protection mechanisms, a review of bankruptcy and corporate recovery laws, and the restructuring of public debt at all levels of government. They also included important industry-specific elements, with new regulatory frameworks and changes to infrastructure business models, including concession mechanisms, de-nationalization and the restructuring of government-owned entities. Keywords: Infrastructure Assets, Financing Risk, Information Asymmetry. The 1994 prices stabilization and the 2002 political transition played a key role in the Brazilian economy’s transformation. The former eliminated the main source of long-term economic unpredictability. The latter significantly reduced the political risks that had previously inhibited long-term investments or made them significantly more expensive. On the macroeconomic level at least, the pillars that ensured economic stability and policy continuation were maintained for almost ten years, despite some deterioration, starting in 2003, in terms of regulatory independence and the promotion of competition within industries. The fact that the world experienced a period of ample supply of funds for investment in emerging countries, in different moments, is relevant for infrastructure. Firstly, the 2000s were a period of remarkable growth worldwide, with re- latively risk perception around the globe. After the 2008 financial crisis in the US and the subsequent European crisis, the expansionist monetary policy in developed countries led to funds overflowing to emerging economies as a side effect, in a movement that led to the generalized appreciation of currencies operating on flexible foreign exchange regimens. Brazil was an important beneficiary of this flow of funds, as it enjoyed, at the time, significant global attention as a result of the broad reforms had in the previous period and the smooth political transition, as discussed earlier. This array of institutional changes and the global environment significantly raised not only the potential for investment in Brazil’s infrastructure but also credit supply and private-sector financing (both domestic and international) to the industries involved. However, in spite of the extraor- 41 dinary window of opportunity and the announcement of a major Federal-Government plan, Brazilian infrastructure failed to take off as expected. The opportunities were therefore under-exploited. There are at least two reasons for the performance to prove itself dissatisfactory despite such a highly propitious environment. The first reason is related to the very nature of infrastructure assets: projects in this area require massive volumes of capital and would face significant financing challenges in any country. Such challenges are magnified in an economy such as Brazil’s. The second reason concerns the increase in political risk since the past decade, due to greater political intervention in technical decisions, which again made projects more expensive and reduced investor propensity to accept them. This article discusses the Brazilian infrastructure-financing environment, as well as its main challenges. Section one addresses the typical characteristics of infrastructure assets, which are relevant to their financing. We then address the old state-support Vs. marketplace dichotomy, which is strongly associated with the role of the National Bank for Economic and Social Development (BNDES) and other institutions that have historically concentrated the supply of funds in this area. We next provide comments on the financial and capital markets and then evaluate infrastructure financing risks. The final section highlights some of the most important challenges remaining in the arena. Characteristics of Infrastructure Assets Generally speaking, infrastructure assets have specific characteristics that vary significantly among industries – power, transport, sanitation and telecommunications – and also within each industry. Depending on an activity’s position on the industry chain, the characteristics may vary significantly. On the power chain, for example, cable segments (transmission and distribution) have marked natural-monopoly characteristics that require strict regulation, while the energy segments (generation and commercialization) operate in more competitive environments. The table next summarizes characteristics of the various infrastructure industries. ■ Table 1: Characteristics of the various infrastructure industries. Consequence Typical response High fixed cost Natural monopoly issues Public operation or concessions; regulation Specific assets Discourages investment Long-maturity contracts Positive externalities Public policy characteristics Access and universalization policies Long life Operator-change issues Operator selection, renewal and replacement processes Stable demand Defensive assets Source: developed by the authors 42 One typical response to these characteristics is the heavy weight of public-sector operation and, even in the case of private or partnered operation, the financing of infrastructure projects shows high public-sector concentration worldwide. Public operation is being increasingly replaced by an increase in various forms of private-sector participation. In Brail, this is seen mainly in contracts under the Concessions Act (Law No. 8.987 of 1995) and the Public-Private Partnerships Act (Law No. 11.079 of e 2004), among other mechanisms. In recent decades, there has been a strong global trend of implementing institutional risk-mitigation mechanisms, such as independent regulation and long-term planning. These enable significantly encouraging not only private investment, but also private-sector financing of infrastructure projects. Brazil, too, has made significant headway in this direction, with the definition of industry frameworks and relatively autonomous regulatory entities, all of which face the dual challenge of being on the first phase of the institutional learning curve and being under constant attack from politicians interested in decisions of high economic and electoral worth. In addition to institutional issues, the financing system has several specific aspects of is own. To illustrate, in the presence of strong information asymmetry for the extension of financing, a common form of defense for finance providers often materializes as requiring warranties that may be real (mortgage, chattel mortgage, commercial pledge, or collateral depo- sit) or personal (sureties and bonds), with the frequent accumulation of either. The difficulty or cost involved in obtaining such warranties may cause companies with viable projects and the willingness to repay financing to lose access to funding. BNDES itself requires high warranties. Given this context, warranties become an important competitiveness-inhibiting factor for several important industries, in addition to inhibiting investments towards universalization. State-support vs. the Marketplace: the Old Argument In brazil and many other countries, infrastructure credit is highly concentrated on public-sector development sources. In the power industry, for example, according to Sabattini (2012, p. 121), development entities – BNDES and Banco do Brasil in particular – are the main individual credits of the majority of conglomerates active in this industry in Brazil. In sanitation, Caixa Econômica Federal, succeeds the former Banco Nacional da Habitação in representing a significant portion of the interest-bearing funds made available to the industry. Several development entities and official institutions provide infrastructure in Brazil: Banco Nacional de Desenvolvimento Econômico e Social (BNDES), Caixa Econômica Federal (CEF), Banco do Brasil (BB), Banco do Nordeste do Brasil (BNB), Banco da Amazônia (BASA), Financiadora de Estudos e Projetos (FINEP), and state-level development agencies, which ope- 43 rate state-level funds as well as pass-throughs from BNDES itself. In addition, there are international development entities, which include, chiefly the World Bank, the International Finance Corporation (IFC), the Inter-American Development Bank (IADB), Corporación Andina de Fomento (CAF), to name a few. Given the importance of infrastructure for economic development and the volume of funds involved, governments are typically directly involved in the financing process. In many cases, however, they fail to ask a crucial question: to what point to development and market-based credit products are substitute or complementary goods? On other words, when governments play a direct role in the financing process, are they simply crowding out private-sector financing and thereby unnecessarily transferring risks and any subsidies to the taxpayer? This question, although of great economic and social import, is also very difficult to answer. The Financial and Capital Markets The financial and capital markets fund infrastructure projects in several ways, with both debt and equity. Financing environments include corporate finance, which creates liabilities directly on the service provider organizations’ balance sheets, and project finance, which is intended for Specific Purpose Entities (“Sociedades de Propósito Específico” – SPE). The corporate mode is more traditionally used in Brazilian infrastructure, particularly due to the massive state-owned companies that prevailed in the operational arena for decades. In this context, own funds generation was an important investment finance instrument, as seen in the economy’s various industries in an environment of high macroeconomic instability and low capital markets development. Under the Planasa basic-sanitation system for example, public-sector financing was provided by the Banco Nacional da Habitação (BNH), together with the state governments that controlled the operational companies, in the belief that such companies would become self-sustainable and start generating the funds to sustain their future expansion and operation. However, in the 1970s and ‘80s, in several infrastructure segments, utility bills started to be used as macroeconomic policy instruments in an attempt to keep inflation in check, sacrificing internally-generated financing. More recently, the Brazilian marketplace’s institutional development and the new Public-Private Partnership instruments of recent decades have incremented the role of project financing. Even major infrastructure operators have been engaging in SPE-based financing schemes. One important program concerns the development of Infrastructure Debentures and Receivables Investment Funds (“Fundos de Investimento em Direitos Creditórios” – FIDC), Incentivized Debenture Funds, and Infrastructure Investment Participation Funds (“Fundos de Investimentos em Participação em Infraestrutura” – FIP- 44 -IE), which rely on tax benefits. The private equity and venture capital (PE/VC) industry is often mentioned as a potential investor in infrastructure. But opinions vary as to the feasibility of investment from this source. For one thing, the typical PE/VC investment cycle – with exit in 3-5 years – does not match the lengthy return periods of typical infrastructure projects. The same critics of the role of private equity and venture capital funds mention that the returns on infrastructure projects ten to be controlled, at least ex-ante, by the public sector by means of the concession process models. To obtain higher returns, investors usually seek out environments with less institutional security, which may disproportionally increase risks, reducing attractiveness. However, one may argue that these characteristics do not imply a natural incompatibility between PE/VC funds and the infrastructure industry, nor do they severely limit the kinds of projects the industry might be willing to finance within the context of infrastructure. In this sense, private equity and venture capital expert Leonardo de Lima Ribeiro recently noted that some of the best PE/VC investments in Latin America were done precisely in infrastructure or related industries, such as: Mills, Cemar, GasAtacama, CPFL, Gol, ALL, among others. For Ribeiro: “the secret lies in investing mainly in the companies that develop projects in the industry, and not necessarily in the projects themselves. And to seek out turnaround opportunities, when the initial investment has already been made but at some point the company got into debt and had to be sold by the original investor. Or privatizations. In fact, with the recent change in power industry bills, I foresee a new round of investments ready to take place soon. Also on the power industry, it is perhaps the segment with the greatest opportunity for private equity investment directly in projects. After all, you can sell power in the long-term via PPAs and get credit to build the developments, which greatly leverages the return on equity. You can also exploit variations in power prices (enter into PPAs when prices are high and purchase power from the market at a lower price) although this has led to the demise of at least one company, a Campanario/Tierra Amarilla). Finally, power includes the issue of environmental risk. You can use equity to develop a project and sell it after securing the required licenses” (Mr. Leonardo de Lima Ribeiro, in an exclusive statement for this article, April 19th, 2013). Financing in the Partnerships Arena Partnerships between the public and private sectors are among the most important vehicles to enable new infrastructure investments, particularly given fiscal and budgetary constraints. In Brazil, the law that governs public-private partnerships includes classes that may be regarded as ordinary, administrative, or sponsored concessions. In the former case, Law No. 8.987 of 1995 applies; for the latter two, Law No. 11.079 of 2004. The two complement one another and are in turn complemented by the Bidding Act 45 (Law NO. 8.666 of 93). Generally speaking, the former two address self-sustainable projects with financial flows arising from billings, whereas the law that became known as the “PPP Act” is concerned with project where public funds are required. Finally, there is the overall legal regimen, the discussion about legal tradition and the capital markets, and a large number of aspects associated with the credit market and bankruptcy laws. In the Brazilian case, in practice, the framework for public-public and public-private contracts has been enabling a significant increase in the range of possibilities in terms of the provision and financing of infrastructure services, allowing new forms of cooperation among a wide array of public and private actors. As an example of partnership and financing institutions in the Brazilian case, the Graphic 1 below illustrates a hypothetical infrastructure projects portfolio. The vertical axis represents the projects’ private rate of return; the horizontal axis shows their social rate of return, which represents the volume of externalities generated by unit of investment. Each sphere represents one project, where the volume of the sphere may be taken as the relative size of the project. The best projects are those in the Northeasternmost part of the first quadrant, as they represent a fortunate combination of high private return and high social return. Sadly, projects like these, which do not require public funds and add great collective benefits, are rare. They are candidates for operation ■ Graphic 1: A hypothetical infrastructure projects portfolio. Source: developed by the authors 46 via ordinary concession, under Law No. 8.987 of 95, or Concessions Act. The projects that lie close to the horizontal axis and to the East of the vertical axis show good social return but low private return. For these, the typical instrument is Law No. 11.079 of 2004 (Public-Private Partnerships Act). These contracts ma also operate with public subsidies, as, despite their social interest, they do not offer sufficient private-sector return to stimulate autonomous execution by private investors. Law No. 12.766 of 2012 (resulting from the congressional approval of Provisional Decree No. 575) created the latest PPP-related developments with the establishment of public investment in projects, which is relevant to project financing insofar as it avoids the need to obtain financing for investments. The Law was regulated by Federal Revenue Service Instruction No. 1.342 which provides special tax treatments for public-entity investments inPPPs. Infrastructure Credit Risks It is worth pointing out, however, that the kinds of risks taken in the infrastructure area may differ from those in more traditional commercial areas. For example, public-sector risk is usually higher, subject to adverse decisions involving regulation, billing, technical standards and other. These characteristics may require a more specialized research structure and more intensive legal advice to protect against risks, for example via analysis of the concessions’ insurance and warranties schedule, or more intense survey of local, idiosyncratic characteristics of the sub-national governments and political environments. Risks in the infrastructure industry are directly related to the institutional environment, both macro- – nationally and locally – and microeconomically. In terms of the national macroeconomic environment, there have been significant gains for infrastructure projects in the past two decades. These changes include addressing the needs for public-sector financing, in the late 1990s, by means of a set of reforms that included the refinancing of sub-national debt, the Fiscal Stability Program, and the Fiscal Responsibility Act; the consolidation of price stability after severe tests involving emerging-economy crises; and the reduction of political risk as an important consequence of the continued economic policy regimen after the 2002 transition. The achievement of investment grade was associated with these factors, which brought about not only reduced public and foreign financing needs, but also increased financing security, with a drop in the perceived probability of solvency crises. Regulatory aspects have also improved significantly in Brazil in recent decades, especially in the latter half of the 1990s and the early 2000s. The creation of regulatory frameworks based on self-governed entities with independent mandates is a true structural departure from the preceding period’s practice. This independence, however, has proved itself fragile in practice, given the political interest in 47 maintaining decisions of high electoral and economic worth under the direct control of the Executive Branch and politically appointed cabinet members. There has also been an important evolution – especially at the state level – in terms of projects with private involvement and governed by the Concessions and Public-Private Partnership (PPP) Acts. Among the important developments, we may name the concessions’ risk matrices, which have been covering important risks that are now more clearly established in the projects. Infrastructure Financing Challenges Based on the above discussion, certain relevant challenges stand out facing the expansion of infrastructure financing in Brazil. Next, we offer a non-exhaustive discussion of these challenges. Firstly, we must evaluate the cost-to-benefit ratio involved in taking major fiscal risks as a result of the public financing mechanisms now available. The public sector’s sustainability is a global challenge. Secondly, we must determine whether public- and private-sector financings are substitute or complementary goods. It is possible that development-agency funding effectively displaces private-sector capital, as some suggest, which would require clearly urgent in-depth studies. Naturally, if they are substitute goods, they will create fiscal risks and bring about undesi- rable effects for the economy. It is therefore worth reviewing development-agency funding in a holistic manner, as well as the role of bank credit and of the capital markets in the Brazilian infrastructure financing. Thirdly, achieving greater leverage, that is, a larger share of credit relative to equity in infrastructure projects. Typically, this involves more intense use of project finance-related instruments, including concessions and public-private partnerships. The country has already made significant legislative progress in this sense, but there is still much to do. Fourth, we must review the institutional environment as concerns the most relevant industry risks, in particular relative to stability and the introduction of competition. The stable macroeconomic environment that has prevailed since the late 1990s and proved itself clearly favorable to investments is dwindling as a result of more discretionary and short-term oriented policies. Likewise, the industry regulation environment is increasingly subject to interventions, in many cases devoid of long-term planning and occasionally aligned with an electoral agenda. Promoting competition wherever possible, embracing the best elements of international experience, would enable safe gains for users. These are complex, inter-related challenges that involve crucial answers for Brazil’s development. Authors 48 Frederico Araujo Turolla Frederico Araujo Turolla is a Doctor of Corporate Economics, a Professor of the ESPM International Management Master’s Program, and a partner at Pezco Microanalysis – e-mail [email protected] Márcio Fernandes Gabrielli Márcio Fernandes Gabrielli is a Maser of Business Administration with emphasis in Finance, a Finance Professor at FGV and ESPM, and an associate at Pezco Microanalysis – e-mail [email protected] Igor J. C. Gondim Igor J. C. Gondim is a Doctoral Candidate, Finance, at FGV, and a consultant at Pezco Microanalysis – e-mail Bibliography [email protected] SABBATINI, Rodrigo. Financiamento do investimento no setor de energia elétrica. In: IPEA. Infraestrutura e Planejamento no Brasil: Coordenação estatal da regulação dos incentivos em prol do investimento - o caso do setor elétrico. IPEA, Relatório de Pesquisa, 2012. TUROLLA, Frederico A. Financiamentos e garantias. Revista Conjuntura da Infraestrutura (FGV e ABDIB). Março de 2010. VALENTE, Paulo Gurgel. Financiamento de longo prazo: um roteiro prático para o BNDES, IFC, FINEP e outras instituições. Rio de Janeiro: Campus Elsevier, 2012. Positive Prepare sua empresa para extrair o máximo de benefícios Saiba como explorar todo o potencial dessa nova realidade Conceda mais crédito com menos riscos Aperfeiçoe sua gestão de clientes Minimize prejuízos causados por fraudes e inadimplência dos dados positivos com a ajuda da Serasa Experian. Positive-se! Contrate agora mesmo a Consultoria Serasa Experian em Cadastro Positivo e rentabilize suas decisões de negócios. Para saber mais, acesse serasaexperian.com.br/cadastropositivo ou ligue para 0800 773 7728 Aumente o retorno das suas ações de cobrança + ++ Revista Tecnologia de Crédito online Acesse informação de ponta, onde você estiver. Sua revista Tecnologia de Crédito é publicada exclusivamente na web. É mais um passo da Serasa Experian para a redução do impacto de suas atividades sobre o meio ambiente, além de um alinhamento à tendência global da universalização de conteúdos. Acesse: serasaexperian.com.br/tecnologia-de-credito, e mantenha-se bem informado sobre o mercado de crédito.