File Download
There are no files associated with this item.
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.1016/j.ejor.2025.02.031
- Scopus: eid_2-s2.0-105002677032
- Find via

Supplementary
-
Citations:
- Scopus: 0
- Appears in Collections:
Article: k-Tree: Crossing sharp boundaries in regression trees to find neighbors
| Title | k-Tree: Crossing sharp boundaries in regression trees to find neighbors |
|---|---|
| Authors | |
| Keywords | Adaptive distance metric Decision tree k-nearest neighbors Machine learning |
| Issue Date | 17-Apr-2025 |
| Publisher | Elsevier |
| Citation | European Journal of Operational Research, 2025, v. 324, n. 3, p. 567-579 How to Cite? |
| Abstract | Traditional classification and regression trees (CARTs) utilize a top-down, greedy approach to split the feature space into sharply defined, axis-aligned sub-regions (leaves). Each leaf treats all of the samples therein uniformly during the prediction process, leading to a constant predictor. Although this approach is well known for its interpretability and efficiency, it overlooks the complex local distributions within and across leaves. As the number of features increases, this limitation becomes more pronounced, often resulting in a concentration of samples near the boundaries of the leaves. Such clustering suggests that there is potential in identifying closer neighbors in adjacent leaves, a phenomenon that is unexplored in the literature. Our study addresses this gap by introducing the -Tree methodology, a novel method that extends the search for nearest neighbors beyond a single leaf to include adjacent leaves. This approach has two key innovations: (1) establishing an adjacency relationship between leaves across the tree space and (2) designing novel intra-leaf and inter-leaf distance metrics through an optimization lens, which are tailored to local data distributions within the tree. We explore three implementations of the -Tree methodology: (1) the Post-hoc -Tree (P-Tree), which integrates the -Tree methodology into constructed decision trees, (2) the Advanced -Tree, which seamlessly incorporates the -Tree methodology during the tree construction process, and (3) the P-random forest, which integrates the P-Tree principles with the random forest framework. The results of empirical evaluations conducted on a variety of real-world and synthetic datasets demonstrate that the -Tree methods have greater prediction accuracy over the traditional models. These results highlight the potential of the -Tree methodology in enhancing predictive analytics by providing a deeper insight into the relationships between samples within the tree space. |
| Persistent Identifier | http://hdl.handle.net/10722/369101 |
| ISSN | 2023 Impact Factor: 6.0 2023 SCImago Journal Rankings: 2.321 |
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | Tian, Xuecheng | - |
| dc.contributor.author | Wang, Shuaian | - |
| dc.contributor.author | Zhen, Lu | - |
| dc.contributor.author | Shen, Zuo-Jun Max | - |
| dc.date.accessioned | 2026-01-17T00:35:25Z | - |
| dc.date.available | 2026-01-17T00:35:25Z | - |
| dc.date.issued | 2025-04-17 | - |
| dc.identifier.citation | European Journal of Operational Research, 2025, v. 324, n. 3, p. 567-579 | - |
| dc.identifier.issn | 0377-2217 | - |
| dc.identifier.uri | http://hdl.handle.net/10722/369101 | - |
| dc.description.abstract | <p>Traditional classification and regression trees (CARTs) utilize a top-down, greedy approach to split the feature space into sharply defined, axis-aligned sub-regions (leaves). Each leaf treats all of the samples therein uniformly during the prediction process, leading to a constant predictor. Although this approach is well known for its interpretability and efficiency, it overlooks the complex local distributions within and across leaves. As the number of features increases, this limitation becomes more pronounced, often resulting in a concentration of samples near the boundaries of the leaves. Such clustering suggests that there is potential in identifying closer neighbors in adjacent leaves, a phenomenon that is unexplored in the literature. Our study addresses this gap by introducing the -Tree methodology, a novel method that extends the search for nearest neighbors beyond a single leaf to include adjacent leaves. This approach has two key innovations: (1) establishing an adjacency relationship between leaves across the tree space and (2) designing novel intra-leaf and inter-leaf distance metrics through an optimization lens, which are tailored to local data distributions within the tree. We explore three implementations of the -Tree methodology: (1) the Post-hoc -Tree (P-Tree), which integrates the -Tree methodology into constructed decision trees, (2) the Advanced -Tree, which seamlessly incorporates the -Tree methodology during the tree construction process, and (3) the P-random forest, which integrates the P-Tree principles with the random forest framework. The results of empirical evaluations conducted on a variety of real-world and synthetic datasets demonstrate that the -Tree methods have greater prediction accuracy over the traditional models. These results highlight the potential of the -Tree methodology in enhancing predictive analytics by providing a deeper insight into the relationships between samples within the tree space.<br></p> | - |
| dc.language | eng | - |
| dc.publisher | Elsevier | - |
| dc.relation.ispartof | European Journal of Operational Research | - |
| dc.rights | This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License. | - |
| dc.subject | Adaptive distance metric | - |
| dc.subject | Decision tree | - |
| dc.subject | k-nearest neighbors | - |
| dc.subject | Machine learning | - |
| dc.title | k-Tree: Crossing sharp boundaries in regression trees to find neighbors | - |
| dc.type | Article | - |
| dc.identifier.doi | 10.1016/j.ejor.2025.02.031 | - |
| dc.identifier.scopus | eid_2-s2.0-105002677032 | - |
| dc.identifier.volume | 324 | - |
| dc.identifier.issue | 3 | - |
| dc.identifier.spage | 567 | - |
| dc.identifier.epage | 579 | - |
| dc.identifier.eissn | 1872-6860 | - |
| dc.identifier.issnl | 0377-2217 | - |
