With the dramatic advances in deep learning technology, machine learning research is focusing on improving the interpretability of model predictions as well as prediction performance in both basic and applied research. While deep learning models have much higher prediction performance than traditional machine learning models, the specific prediction process is still difficult to interpret and/or explain. This is known as the black-boxing of machine learning models and is recognized as a particularly important problem in a wide range of research fields, including manufacturing, commerce, robotics, and other industries where the use of such technology has become commonplace, as well as the medical field, where mistakes are not tolerated. This bulletin is based on the summary of the author's dissertation. The research summarized in the dissertation focuses on the attention mechanism, which has been the focus of much attention in recent years, and discusses its potential for both basic research in terms of improving prediction performance and interpretability, and applied research in terms of evaluating it for real-world applications using large data sets beyond the laboratory environment. The dissertation also concludes with a summary of the implications of these findings for subsequent research and future prospects in the field.
翻译:隨著深度學習技術的飛速發展,機器學習研究正致力於提高模型預測性能和可解釋度,在基礎和應用研究中都是如此。儘管深度學習模型的預測性能比傳統機器學習模型高得多,但具體的預測過程仍然很難解釋或解釋。這被稱為機器學習模型的黑盒化,是許多研究領域的一個尤其重要的問題,包括製造、商業、機器人和其他使用此類技術的行業,以及醫療領域,這些地方不容許出錯。本文基於作者的論文摘要。論文研究著眼於注意力機制,這是近年來受到關注的焦點,並討論了它對於基礎研究和應用研究的潛在影響。該論文也總結了這些發現對後續研究和未來展望的影響。