Abstract
Thousands of variables are measured in line during the manufacture of central processing units (cpus). Once the manufacturing process is complete, each chip undergoes a series of tests for functionality that determine the yield of the manufacturing process. Traditional statistical methods such as ANOVA have been used for many years to find relationships between end of line yield and in line variables that can be used to sustain and improve process yield. However, a large increase in the number of variables being measured in line due to modern manufacturing trends has overwhelmed the capability of traditional methods. A filter is needed between the tens of thousands of variables in the database and the traditional methods. In this paper, we propose using true multivariate feature selection capable of dealing with complex, mixed typed data sets as an initial step in yield analysis to reduce the number of variables that receive additional investigation using traditional methods. We demonstrate this approach on a historical data set with over 30,000 variables and successfully isolate the cause of a specific yield problem.
Access provided by Autonomous University of Puebla. Download to read the full chapter text
Chapter PDF
Similar content being viewed by others
References
Yu, L., Liu, H.: Efficient Feature Selection via Analysis of Relevance and Redundancy. J. of Mach. Learn. Res. 5, 1205–1224 (2004)
Koller, D., Sahami, M.: Toward Optimal Feature Selection. In: Proceedings of the Thirteenth International Conferrence on Machine Learning, pp. 284–292. Morgan Kaufmann Publishers, San Francisco (1996)
John, G.H., Kohavi, R., Pfleger, K.: Irrelevant Features and the Subset Selection Problem. In: Machine Learning: Proceedings of the Eleventh International Conference, pp. 121–129. Morgan Kaufmann Publishers, San Francisco (1994)
Tuv, E., Borisov, A., Runger, G., Torkkola, K.: Feature Selection with Ensembles, Artificial Variables, and Redundancy Elimination. J. of Mach. Learn. Res. 10, 1341–1366 (2009)
Breiman, L.: Random Forests. Mach. Learn. 45(1), 5–32 (2001)
Friedman, J.H.: Greedy Function Approximation: A Gradient Boosting Machine. Annals of Statistics 29, 1189–1232 (2001)
Breiman, L., Friedman, J.H., Olshen, R.A., Stone, C.J.: Classification and Regression Trees. Chapman and Hall, New York (1998)
St. Pierre, E.R., Tuv, E., Borisov, A.: Spatial Patterns in Sort Wafer Maps and Identifying Fab Tool Commonalities. In: IEEE/SEMI Advanced Semiconductor Manufacturing Conference, pp. 268–272. IEEE, Los Alamitos (2008)
Borisov, A., Eruhimov, V., Tuv, E.: Dynamic soft feature selection for tree-based ensembles. In: Guyon, I., Gunn, S., Nikravesh, M., Zadeh, L. (eds.) Feature Extraction, Foundations and Applications. Springer, New York (2005)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2011 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
St. Pierre, E., Tuv, E. (2011). Robust, Non-Redundant Feature Selection for Yield Analysis in Semiconductor Manufacturing. In: Perner, P. (eds) Advances in Data Mining. Applications and Theoretical Aspects. ICDM 2011. Lecture Notes in Computer Science(), vol 6870. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-23184-1_16
Download citation
DOI: https://doi.org/10.1007/978-3-642-23184-1_16
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-23183-4
Online ISBN: 978-3-642-23184-1
eBook Packages: Computer ScienceComputer Science (R0)