Groupwise Bayesian Dimension Reduction

2017 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN)(2017)

引用 0|浏览15
暂无评分
摘要
Nearly all existing estimations of the central subspace in regression take the frequentist approach. However, when the predictors fall naturally into a number of groups, these frequentist methods treat all predictors indiscriminately and can result in loss of the group-specific relation between the response and the predictors. In this article, we propose a Bayesian solution for dimension reduction which incorporates such group knowledge. We place a prior whose variance is constrained to the form of a direct sum on the central subspace and directly model the response density in terms of the sufficient predictors using a finite mixture model. This approach is computationally efficient and offers a unified framework to handle categorical predictors, missing predictors, and Bayesian variable selection. We illustrate the method using both a simulation study and an analysis of a temperature data set.
更多
查看译文
关键词
groupwise Bayesian dimension reduction,frequentist approach,group-specific relation,group knowledge,variance,central subspace,finite mixture model,categorical predictors,missing predictors,Bayesian variable selection,temperature data set analysis
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要