J. C. Castro Gómez, E. Carrizosa Priego, V. Guerrero

In many real-world scenarios, data are collected from diverse sources, some of which can be costly to access in terms of resources or time. This often leads to incomplete datasets in the sense that blocks of features are available only for some of the observations, thus yielding a block-missing data set. Rather than discarding or imputing missing data, in this work we propose a greedy tree-based method that makes use of all available data to train a predictive model.

At each step, this greedy algorithm evaluates the option of making a split based on a block of features, aiming to balance predictive accuracy with the cost of accessing additional information. Designed for both classification and regression tasks, our approach integrates an adaptive decision-making mechanism to optimize the trade-off between data completeness and prediction performance.

Keywords: Pescriptive model, Classification and Regression Trees, Random Forest

Scheduled

GT03. AMC1 Machine Learning
June 10, 2025  11:30 AM
Auditorio 1. Ricard Vinyes


Other papers in the same session

Constrained additive models for global optimization

C. D'Ambrosio, M. Cuesta Santa Teresa, M. Durbán, V. Guerrero, R. Spencer Trindade

Medoides para conjuntos difusos p-dimensionales

B. Sinova Fernández, M. Á. Gil Álvarez


Cookie policy

We use cookies in order to be able to identify and authenticate you on the website. They are necessary for the correct functioning of it, and therefore they can not be disabled. If you continue browsing the website, you are agreeing with their acceptance, as well as our Privacy Policy.

Additionally, we use Google Analytics in order to analyze the website traffic. They also use cookies and you can accept or refuse them with the buttons below.

You can read more details about our Cookie Policy and our Privacy Policy.