Papers
Topics
Authors
Recent
Search
2000 character limit reached

Entropy Based Risk Measures

Published 22 Jan 2018 in math.PR | (1801.07220v1)

Abstract: Entropy is a measure of self-information which is used to quantify losses. Entropy was developed in thermodynamics, but is also used to compare probabilities based on their deviating information content. Corresponding model uncertainty is of particular interest in stochastic programming and its applications like mathematical finance, as complete information is not accessible or manageable in general. This paper extends and generalizes the Entropic Value-at-Risk by involving R\'enyi entropies. We provide explicit relations of different entropic risk measures, we elaborate their dual representations and present their relations explicitly. We consider the largest spaces which allow studying the impact of information in detail and it is demonstrated that these do not depend on the information loss. The dual norms and Hahn-Banach functionals are characterized explicitly.

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Authors (2)

Collections

Sign up for free to add this paper to one or more collections.