Gaussian oblique decision tree technique for classification and regression
Gaussian oblique decision tree technique for classification and regression
Decision tree techniques remain a cornerstone of interpretable machine learning techniques. Specifically, axis-parallel decision tree techniques are very popular in real-world applications. However, traditional axis-parallel decision tree techniques that perform orthogonal splits construct staircase decision boundaries, which would limit the predictive power when presented with data that has strong feature interactions. Hence, we propose a novel oblique decision tree technique that constructs a multivariate decision tree, called Gaussian mixture model Oblique Decision Trees (GODT) for classification and regression. We also propose a novel quadratic decision tree technique that constructs quadratic decision boundaries at each non-terminal node to split the data, called Gaussian mixture model Quadratic Decision Tree (GQDT) for classification. In these proposed techniques, the Gaussian mixture model is efficiently incorporated into the architecture of a binary tree-building process. We propose a novel and computationally cheaper way to construct hyperplanes for the oblique tree and hypersurface for the quadratic decision tree techniques. This is because, most of the current techniques have a high computational cost associated to them when trying to achieve a good prediction accuracy. Experiments using the GODT for classification and regression tasks, and GQDT on classification tasks on open-source datasets illustrate that both GODT and GQDT performs better in most cases when compared against existing established techniques both in test accuracy and computational times. Furthermore, as single tree-based techniques are known to have high-variance and sensitivity to training data, we extend our work to build an ensemble of GODT and GQDT techniques. Finally, we run extensive experiments to test the performance of the techniques that we propose on open-source datasets and we compare the performance against established techniques like the Classification and Regression Trees (CART), Support Vector Machines (SVM), Standard Random Forest (RF), Linear Discriminant Analysis (LDA), and Quadratic Discriminant Analysis (QDA).
University of Southampton
Sivanand, Aditya
71bb268c-2c0c-42dd-af77-5c28f0c40c3d
30 March 2026
Sivanand, Aditya
71bb268c-2c0c-42dd-af77-5c28f0c40c3d
Fliege, Joerg
54978787-a271-4f70-8494-3c701c893d98
Dodd, Erengul
b3faed76-f22b-4928-a922-7f0b8439030d
Sivanand, Aditya
(2026)
Gaussian oblique decision tree technique for classification and regression.
University of Southampton, Doctoral Thesis, 176pp.
Record type:
Thesis
(Doctoral)
Abstract
Decision tree techniques remain a cornerstone of interpretable machine learning techniques. Specifically, axis-parallel decision tree techniques are very popular in real-world applications. However, traditional axis-parallel decision tree techniques that perform orthogonal splits construct staircase decision boundaries, which would limit the predictive power when presented with data that has strong feature interactions. Hence, we propose a novel oblique decision tree technique that constructs a multivariate decision tree, called Gaussian mixture model Oblique Decision Trees (GODT) for classification and regression. We also propose a novel quadratic decision tree technique that constructs quadratic decision boundaries at each non-terminal node to split the data, called Gaussian mixture model Quadratic Decision Tree (GQDT) for classification. In these proposed techniques, the Gaussian mixture model is efficiently incorporated into the architecture of a binary tree-building process. We propose a novel and computationally cheaper way to construct hyperplanes for the oblique tree and hypersurface for the quadratic decision tree techniques. This is because, most of the current techniques have a high computational cost associated to them when trying to achieve a good prediction accuracy. Experiments using the GODT for classification and regression tasks, and GQDT on classification tasks on open-source datasets illustrate that both GODT and GQDT performs better in most cases when compared against existing established techniques both in test accuracy and computational times. Furthermore, as single tree-based techniques are known to have high-variance and sensitivity to training data, we extend our work to build an ensemble of GODT and GQDT techniques. Finally, we run extensive experiments to test the performance of the techniques that we propose on open-source datasets and we compare the performance against established techniques like the Classification and Regression Trees (CART), Support Vector Machines (SVM), Standard Random Forest (RF), Linear Discriminant Analysis (LDA), and Quadratic Discriminant Analysis (QDA).
Text
Gaussian Oblique Decision Tree Technique for Classification and Regression
- Version of Record
Text
Final-thesis-submission-Examination-Mr-Aditya-Sivanand
Restricted to Repository staff only
More information
Published date: 30 March 2026
Identifiers
Local EPrints ID: 510392
URI: http://eprints.soton.ac.uk/id/eprint/510392
PURE UUID: a434226e-068d-4d1b-8eea-e2272f5769ec
Catalogue record
Date deposited: 30 Mar 2026 16:41
Last modified: 31 Mar 2026 02:00
Export record
Contributors
Author:
Aditya Sivanand
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics