MyPivots
ForumDaily Notes
Dictionary
Sign In

Risk-Adjusted Return

Risk-adjusted return is a measure of the return on an investment relative to its risk. It is calculated by dividing the investment's return by its risk, as measured by its standard deviation.

The higher the risk-adjusted return, the better the investment is. This is because it means that the investment is generating more return for the amount of risk it is taking on.

There are a number of different ways to calculate risk-adjusted return. The most common method is the Sharpe ratio, which is calculated by dividing the investment's excess return (its return above the risk-free rate) by its standard deviation.

Another common method is the Treynor ratio, which is calculated by dividing the investment's excess return by its beta. Beta is a measure of the investment's volatility relative to the market.

The choice of which risk-adjusted return measure to use depends on the specific investment being evaluated. For example, the Sharpe ratio is often used for evaluating investments with low volatility, while the Treynor ratio is often used for evaluating investments with high volatility.

Risk-adjusted return is an important tool for investors because it allows them to compare investments with different levels of risk and return. By using risk-adjusted return, investors can identify investments that are generating the highest return for the amount of risk they are taking on.

Here are some additional points to consider about risk-adjusted return:

Overall, risk-adjusted return is a valuable tool for investors. However, it is important to understand its limitations and to use it in conjunction with other investment analysis tools.