A Feature Extraction Method Based on Feature Fusion and its Application in the Text-Driven Failure Diagnosis Field.
DOI:
https://doi.org/10.9781/ijimai.2020.11.006Keywords:
Feature Extraction, Feature Fusion, Text Mining, Failure DiagnosisAbstract
As a basic task in NLP (Natural Language Processing), feature extraction directly determines the quality of text clustering and text classification. However, the commonly used TF-IDF (Term Frequency & Inverse Document Frequency) and LDA (Latent Dirichlet Allocation) text feature extraction methods have shortcomings in not considering the text’s context and blindness to the topic of the corpus. This study builds a feature extraction algorithm and application scenarios in the field of failure diagnosis. A text-driven failure diagnosis model is designed to classify and automatically judge which failure mode the failure described in the text belongs to once a failure-description text is entered. To verify the effectiveness of the proposed feature extraction algorithm and failure diagnosis model, a long-term accumulated failure description text of an aircraft maintenance and support system was used as a subject to conduct an empirical study. The final experimental results also show that the proposed feature extraction method can effectively improve the effect of clustering, and the proposed failure diagnosis model achieves high accuracies and low false alarm rates.
Downloads
References
[1] Cambria, E. & White, B., “Jumping NLP Curves: A Review of Natural Language Processing Research [Review Article]”, Computational Intelligence Magazine IEEE, vol. 9, no. 2, pp. 48-57, 2014.
[2] Tom, Y., Devamanyu, H., Soujanya, P. & Erik, C., “Recent Trends in Deep Learning Based Natural Language Processing [Review Article]”, IEEE Computational Intelligence Magazine, vol. 13, no. 3, pp. 55-75, 2018.
[3] Salton, G., Wong, A. & Yang, C., “A vector space model for automatic indexing”, Communications of the ACM, vol. 18, no. 11, pp. 613-620, 1975.
[4] Liang, H., Sun, X., Sun, Y. & Gao, Y., “Text feature extraction based on deep learning: a review”, EURASIP journal on wireless communications and networking, vol. 2017, no. 1, pp. 1-12, 2017.
[5] Sparck Jones, K., “A statistical interpretation of term specificity and its application in retrieval”, Journal of documentation, vol. 28, no. 1, pp. 11- 21, 1972.
[6] Blei, D. M., Ng, A. Y. & Jordan, M. I., “Latent dirichlet allocation”, Journal of machine Learning research, vol. 3, no. Jan, pp. 993-1022, 2003.
[7] Wang, F., Xu, T., Tang, T., Zhou, M. & Wang, H., “Bilevel feature extraction-based text mining for fault diagnosis of railway systems”, IEEE transactions on intelligent transportation systems, vol. 18, no. 1, pp. 49-58, 2016.
[8] Rajpathak, D. G. & Singh, S., “An Ontology-Based Text Mining Method to Develop D-Matrix From Unstructured Text”, IEEE Transactions on Systems Man & Cybernetics Systems, vol. 44, no. 7, pp. 966-977, 2014.
[9] Rodrigues, R. E. R. S., Balestrassi, P. P., Paiva, A. P., Garcia-Diaz, A. & Pontes, F. J., “Aircraft interior failure pattern recognition utilizing text mining and neural networks”, Journal of Intelligent Information Systems, vol. 38, no. 3, pp. 741-766, 2012.
[10] Harris, Z. S., “Distributional structure”, Word, vol. 10, no. 2-3, pp. 146- 162, 1954.
[11] Deerwester, S., Dumais, S. T., Furnas, G. W., Landauer, T. K. & Harshman, R., “Indexing by latent semantic analysis”, Journal of the American society for information science, vol. 41, no. 6, pp. 391-407, 1990.
[12] Roweis, S. T. & Saul, L. K., “Nonlinear dimensionality reduction by locally linear embedding”, science, vol. 290, no. 5500, pp. 2323-2326, 2000.
[13] Bengio, Y., Ducharme, R. E. J., Vincent, P. & Jauvin, C., “A neural probabilistic language model”, Journal of machine learning research, vol. 3, no. Feb, pp. 1137-1155, 2003.
[14] Ming, T., Lei, Z., Xianchun, Z. & Others, “Document vector representation based on Word2Vec”, Computer Science, vol. 43, no. 6, pp. 214-217, 2016.
[15] Adali, T. U. L., Levin-Schwartz, Y. & Calhoun, V. D., “Multimodal data fusion using source separation: Application to medical imaging”, Proceedings of the IEEE, vol. 103, no. 9, pp. 1494-1506, 2015.
[16] Jing, L., Wang, T., Zhao, M. & Wang, P., “An adaptive multi-sensor data fusion method based on deep convolutional neural networks for fault diagnosis of planetary gearbox”, Sensors, vol. 17, no. 2, pp. 414, 2017.
[17] Wang, S., Deng, Z. & Yin, G., “An accurate GPS-IMU/DR data fusion method for driverless car based on a set of predictive models and grid constraints”, Sensors, vol. 16, no. 3, pp. 280, 2016.
[18] Khaleghi, B., Khamis, A., Karray, F. O. & Razavi, S. N., “Multisensor data fusion: A review of the state-of-the-art”, Information fusion, vol. 14, no. 1, pp. 28-44, 2013.
[19] Yang, J., Yang, J., Zhang, D. & Lu, J., “Feature fusion: parallel strategy vs. serial strategy”, Pattern recognition, vol. 36, no. 6, pp. 1369-1381, 2003.
[20] Liu, C. & Wechsler, H., “A shape-and texture-based enhanced Fisher classifier for face recognition”, IEEE transactions on image processing, vol. 10, no. 4, pp. 598-608, 2001.
[21] Sun, Q., Zeng, S., Liu, Y., Heng, P. & Xia, D., “A new method of feature fusion and its application in image recognition”, Pattern Recognition, vol. 38, no. 12, pp. 2437-2448, 2005.
[22] Geman, S. & Geman, D., “Stochastic relaxation, Gibbs distributions, and the Bayesian restoration of images”, IEEE Transactions on pattern analysis and machine intelligence, no. 6, pp. 721-741, 1984.
[23] Visser, E., Nijhuis, E. H., Buitelaar, J. K. & Zwiers, M. P., “Partition-based mass clustering of tractography streamlines”, NeuroImage, vol. 54, no. 1, pp. 303-312, 2011.
[24] Bouguettaya, A., Yu, Q., Liu, X., Zhou, X. & Song, A., “Efficient agglomerative hierarchical clustering”, Expert Systems with Applications, vol. 42, no. 5, pp. 2785-2797, 2015.
[25] Tang, X. & Zhu, P., “Hierarchical clustering problems and analysis of fuzzy proximity relation on granular space”, IEEE Transactions on Fuzzy Systems, vol. 21, no. 5, pp. 814-824, 2012.
[26] Lu, J. & Zhu, Q., “An effective algorithm based on density clustering framework”, Ieee Access, vol. 5,4991-5000, 2017.
[27] Tu, L. & Chen, Y., “Stream data clustering based on grid density and attraction”, ACM Transactions on Knowledge Discovery from Data (TKDD), vol. 3, no. 3, pp. 1-27, 2009.
[28] Rodriguez, A. & Laio, A., “Clustering by fast search and find of density peaks”, Science, vol. 344, no. 6191, pp. 1492-1496, 2014.
[29] Jain, A. K., “Data clustering: 50 years beyond K-means”, Pattern recognition letters, vol. 31, no. 8, pp. 651-666, 2010.
[30] Japkowicz, N. & Stephen, S., “The class imbalance problem: A systematic study”, Intelligent data analysis, vol. 6, no. 5, pp. 429-449, 2002.
[31] Chawla, N. V., Bowyer, K. W., Hall, L. O. & Kegelmeyer, W. P., “SMOTE: synthetic minority over-sampling technique”, Journal of artificial intelligence research, vol. 16,321-357, 2002.
[32] Cortes, C. & Vapnik, V., “Support-vector networks”, Machine learning, vol. 20, no. 3, pp. 273-297, 1995.
Downloads
Published
-
Abstract228
-
PDF44






