18 résultats
pour « bayesian »
#bayesian data imputation holds significant importance in a variety of fields including #riskmanagement. Incomplete or missing data can hinder a thorough analysis of risks, making accurate decision-making challenging. By employing imputation techniques to fill in the gaps, risk managers can obtain a more comprehensive and reliable understanding of the underlying risk factors. This, in turn, enables them to make informed decisions and develop effective strategies for #riskmitigation.
"This paper examines a #stochastic one-period #insurancemarket with incomplete information. The aggregate amount of #claims follows a compound #poisson distribution. #insurers are assumed to be exponential utility maximizers, with their degree of #riskaversion forming their private information. A premium strategy is defined as a map between risk types and premium rates. The optimal premium strategies are denoted by the pure-strategy #bayesian #nash equilibrium, whose existence and uniqueness are demonstrated under specific conditions for the demand function..."
This study examines interpersonal heterogeneity in #risk attitudes in #decisionmaking experiments. The use of #bayesian and classical methods for estimating the hierarchical model has sparked debate. Both approaches use the population distribution of risk attitudes to identify individual-specific risk attitudes. Comparing existing experimental data, both methods yield similar conclusions about risk attitudes.
#bayesian data imputation is a technique used to fill in missing data in a variety of fields, including #riskmanagement. By employing imputation techniques to fill in the gaps, #riskmanagers can obtain a more comprehensive and reliable understanding of the underlying #risk factors, enabling them to make informed decisions and develop effective strategies for #riskmitigation.
"... we find that the #bayesian approach outperforms the classical [#counterparty #risk #model] in identifying whether a model is correctly specified, which is the principal aim of any backtesting framework."
Traditional #statistical and #algorithm-based methods used to analyze #bigdata often overlook small but significant evidence. #bayesian #statistics, driven by #conditional #probability, offer a solution to this challenge. The review identifies two main applications of Bayesian statistics in #finance: prediction in financial markets and credit risk models. The findings aim to provide valuable insights for researchers aiming to incorporate Bayesian methods and address the sample size issue effectively in #financial #research.
We offer a #datadriven theory of #belief formation that explains sudden surges in economic #uncertainty and their consequences. It argues that people, like #bayesian econometricians, estimate a distribution of macroeconomic outcomes but do not know the true distribution. The paper shows how real-time estimation of distributions with non-normal tails can result in large fluctuations in uncertainty, particularly related to tail events or "black swans." Using real-time GDP data, the authors find that revisions in estimated #blackswan #risk explain most of the fluctuations in uncertainty. These findings highlight the importance of #accounting for the effects of uncertainty and non-normality in economic decision-making and #policymaking.
This paper focuses on predicting #corporate #default #risk using frailty correlated default #models with subjective judgments. The study uses a #bayesian approach with the Particle Markov Chain #montecarlo algorithm to analyze data from #us public non-financial firms between 1980 and 2019. The findings suggest that the volatility and mean reversion of the hidden factor have a significant impact on the default intensities of the firms.
"In this paper we propose efficient #bayesian Hamiltonian #montecarlo method for estimation of #systemicrisk#measures , LRMES, SRISK and ΔCoVaR, and apply it for thirty global systemically important banks and for eighteen largest #us#financialinstitutions over the period of 2000-2020. The systemic risk measures are computed based on the Dynamic Conditional Correlations model with generalized asymmetric #volatility. A policymaker may choose to rank the firms using some quantile of their systemic risk distributions such as 90, 95, or 99% depending on #risk preferences with higher quantiles being more conservative."
This paper proposes a novel mixed-frequency quantile vector autoregression (MF-QVAR) model that uses a #bayesian framework and multivariate asymmetric Laplace distribution to estimate missing low-frequency variables at higher frequencies. The proposed method allows for timely policy interventions by analyzing conditional quantiles for multiple variables of interest and deriving quantile-related #riskmeasures at high frequency. The model is applied to the US economy to #nowcast conditional quantiles of #gdp, providing insight into #var, Expected Shortfall, and distance among percentiles of real GDP nowcasts.