<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><title>XGBoost on Øystein Sørensen</title><link>https://osorensen.github.io/tags/xgboost/</link><description>Recent content in XGBoost on Øystein Sørensen</description><generator>Hugo -- 0.128.0</generator><language>en</language><lastBuildDate>Wed, 22 Oct 2025 00:00:00 +0000</lastBuildDate><atom:link href="https://osorensen.github.io/tags/xgboost/index.xml" rel="self" type="application/rss+xml"/><item><title>Gaining Brain Insights by Tapping into the Black Box: Linking Structural MRI Features to Age and Cognition using Shapley-Based Interpretation Methods</title><link>https://osorensen.github.io/papers/paper13/</link><pubDate>Wed, 22 Oct 2025 00:00:00 +0000</pubDate><guid>https://osorensen.github.io/papers/paper13/</guid><description>This paper evaluates multiple interpretability techniques for machine learning models applied to neuroimaging data, including SHAP and SAGE. We trained XGBoost models to predict age and fluid intelligence using UK Biobank data and found that subcortical mean intensities are associated with brain aging, while fluid intelligence prediction is driven by the hippocampus and cerebellum. Published in Neuroinformatics.</description></item></channel></rss>