Skip to content

Commit

Permalink
Update 6-uncertainty-overview.md
Browse files Browse the repository at this point in the history
  • Loading branch information
qualiaMachine authored Dec 19, 2024
1 parent d572932 commit d3d9bf0
Showing 1 changed file with 3 additions and 29 deletions.
32 changes: 3 additions & 29 deletions episodes/6-uncertainty-overview.md
Original file line number Diff line number Diff line change
Expand Up @@ -23,10 +23,10 @@ exercises: 0
- Summarize when and where different uncertainty estimation methods are most useful.

::::::::::::::::::::::::::::::::::::::::::::::::
### How confident is my model? Will it generalize to new data or subpopulations?
## How confident is my model? Will it generalize to new data?
Understanding how confident a model is in its predictions is a valuable tool for building trustworthy AI systems, especially in high-stakes settings like healthcare or autonomous vehicles. Model uncertainty estimation focuses on quantifying the model's confidence and is often used to identify predictions that require further review or caution.

### Sources of uncertainty
## Sources of uncertainty
At its core, model uncertainty starts with the **data** itself, as all models learn to form embeddings (feature representations) of the data. Uncertainty in the data—whether from inherent randomness or insufficient coverage—propagates through the model's embeddings, leading to uncertainty in the outputs.

#### 1) Aleatoric (Random) uncertainty
Expand Down Expand Up @@ -98,7 +98,7 @@ For example, in autonomous vehicles, OOD detection can help flag unexpected scen

:::: challenge

#### Identify aleatoric and epistemic uncertainty
## Identify aleatoric and epistemic uncertainty

For each scenario below, identify the sources of **aleatoric** and **epistemic** uncertainty. Provide specific examples based on the context of the application.

Expand Down Expand Up @@ -129,36 +129,10 @@ For each scenario below, identify the sources of **aleatoric** and **epistemic**


#### Summary

Uncertainty estimation is a critical component of building reliable and trustworthy machine learning models, especially in high-stakes applications. By understanding the distinction between aleatoric uncertainty (inherent data noise) and epistemic uncertainty (gaps in the model's knowledge), practitioners can adopt tailored strategies to improve model robustness and interpretability.

- Aleatoric uncertainty is irreducible noise in the data itself. Addressing this requires models that can predict variability, such as heteroscedastic loss functions, or strategies like data augmentation to make models more resilient to imperfections.
- Epistemic uncertainty arises from the model's incomplete understanding of the data distribution. It can be mitigated through methods like Monte Carlo dropout, Bayesian neural networks, ensemble models, and Out-of-Distribution (OOD) detection. Among these methods, OOD detection has become a cornerstone for handling epistemic uncertainty in practical applications. Its ability to flag anomalous or out-of-distribution inputs makes it an essential tool for ensuring model predictions are reliable in real-world scenarios.
- In many cases, collecting more data and employing active learning can directly address the root causes of epistemic uncertainty.

When choosing a method, it’s important to consider the trade-offs in computational cost, model complexity, and the type of uncertainty being addressed. Together, these techniques form a powerful toolbox, enabling models to better navigate uncertainty and maintain trustworthiness in dynamic environments. By combining these approaches strategically, practitioners can ensure that their systems are not only accurate but also robust, interpretable, and adaptable to the challenges of real-world data.

#### References and methods

#### Methods for addressing aleatoric uncertainty
1. **Heteroscedastic models**: Predict noise levels directly using specialized loss functions.


2. **Data augmentation**: Add noise to inputs to improve robustness to imaging artifacts.


#### Methods for addressing epistemic uncertainty
1. **Ensemble models**: Aggregate predictions to capture model disagreement.
- Example application: Predicting rare pandemic impacts on hospital demand.
- Reference: Leibig, C., et al. (2017). "Leveraging uncertainty information from deep neural networks for disease detection."
[Scientific Reports](https://www.nature.com/articles/s41598-017-17876-z).

2. **Bayesian neural networks**: Model uncertainty in parameters for underrepresented scenarios.
- Example application: Detecting rare tumor types in radiology.
- Reference: Blundell, C., et al. (2015). "Weight uncertainty in neural networks."
[ArXiv](https://arxiv.org/abs/1505.05424).

3. **Out-of-distribution detection**: Identify inputs outside the training distribution.
- Example application: Flagging out-of-scope queries in chatbot systems.
- Reference: Hendrycks, D., & Gimpel, K. (2017). "A baseline for detecting misclassified and out-of-distribution examples in neural networks."
[ArXiv](https://arxiv.org/abs/1610.02136).

0 comments on commit d3d9bf0

Please sign in to comment.