TY - GEN
T1 - Cutting down high dimensional data with Fuzzy weighted forests (FWF)
AU - Wang, Tao
AU - Gault, Richard
AU - Greer, Desmond
PY - 2022/9/14
Y1 - 2022/9/14
N2 - Takagi-Sugeno-Kang (TSK) rule-based fuzzy systems struggle to deal with high dimensional data and suffer from the curse of dimensionality. As the number of input features increases, the number of rules increases exponentially, which reduces the model's interpretability rapidly. This paper presents a novel fuzzy weighted forest aggregation method to effectively model high dimensional data by reducing the number of fuzzy rules, without sacrificing accuracy. The fuzzy weighted forest is comprised of several fuzzy weighted trees. Each tree is created based on a subset of features captured across different parts of the input space. Given n input features and N samples, every fuzzy tree is assigned randomly n' features and N' samples, where n' and N' are significantly smaller than n and N respectively. Each path within a tree, from root to leaf, forms a fuzzy rule. The non-leaf nodes represent the antecedents of rules and the leaf node represent the consequents. This study shows how the proposed method utilizes pruning to significantly reduce the number of fuzzy rules. This method therefore creates a less complex model whilst achieving high accuracy comparable with, and sometimes better than, existing state-of-the-art TSK based fuzzy models.
AB - Takagi-Sugeno-Kang (TSK) rule-based fuzzy systems struggle to deal with high dimensional data and suffer from the curse of dimensionality. As the number of input features increases, the number of rules increases exponentially, which reduces the model's interpretability rapidly. This paper presents a novel fuzzy weighted forest aggregation method to effectively model high dimensional data by reducing the number of fuzzy rules, without sacrificing accuracy. The fuzzy weighted forest is comprised of several fuzzy weighted trees. Each tree is created based on a subset of features captured across different parts of the input space. Given n input features and N samples, every fuzzy tree is assigned randomly n' features and N' samples, where n' and N' are significantly smaller than n and N respectively. Each path within a tree, from root to leaf, forms a fuzzy rule. The non-leaf nodes represent the antecedents of rules and the leaf node represent the consequents. This study shows how the proposed method utilizes pruning to significantly reduce the number of fuzzy rules. This method therefore creates a less complex model whilst achieving high accuracy comparable with, and sometimes better than, existing state-of-the-art TSK based fuzzy models.
U2 - 10.1109/FUZZ-IEEE55066.2022.9882660
DO - 10.1109/FUZZ-IEEE55066.2022.9882660
M3 - Conference contribution
T3 - IEEE International Conference on Fuzzy Systems (FUZZ-IEEE): Proceedings
BT - IEEE International Conference on Fuzzy Systems (FUZZ-IEEE): Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
ER -