| 设计 任务书 文档 开题 答辩 说明书 格式 模板 外文 翻译 范文 资料 作品 文献 课程 实习 指导 调研 下载 网络教育 计算机 网站 网页 小程序 商城 购物 订餐 电影 安卓 Android Html Html5 SSM SSH Python 爬虫 大数据 管理系统 图书 校园网 考试 选题 网络安全 推荐系统 机械 模具 夹具 自动化 数控 车床 汽车 故障 诊断 电机 建模 机械手 去壳机 千斤顶 变速器 减速器 图纸 电气 变电站 电子 Stm32 单片机 物联网 监控 密码锁 Plc 组态 控制 智能 Matlab 土木 建筑 结构 框架 教学楼 住宅楼 造价 施工 办公楼 给水 排水 桥梁 刚构桥 水利 重力坝 水库 采矿 环境 化工 固废 工厂 视觉传达 室内设计 产品设计 电子商务 物流 盈利 案例 分析 评估 报告 营销 报销 会计 | |||||
|
|||||
|
|||||
|
|||||
The MNIST Database of Handwritten Digit Images for Machine Learning Research
In this issue, ЁАBest of the WebЁБ presents the modified National Institute of Standards and Technology (MNIST) resources, consisting of a collection of handwritten digit images used extensively in optical character recognition and machine learning research.
Handwritten digit recognition is an important problem in optical character recognition, and it has been used as a test case for theories of pattern recognition and machine learning algorithms for many years. Historically, to promote machine learning and pattern recognition research, several standard databases have emerged in which the handwritten digits are preprocessed, including segmentation and normalization, so that researchers can compare recognition results of their techniques on a common basis. The freely available MNIST database of handwritten digits has become a standard for fast-testing machine learning algorithms for this purpose. The simplicity of this task is analogous to the TIDigit (a speech database created by Texas Instruments) task in speech recognition. Just like there is a long list for more complex speech recognition tasks, there are many more difficult and challenging tasks for image recognition and computer vision, which will not be addressed in this column.
DATA
The MNIST database was constructed out of the original NIST database; hence, modified NIST or MNIST. There are 60,000 training images (some of these training images can also be used for crossvalidation purposes) and 10,000 test images, both drawn from the same distribution. All these black and white digits are size normalized, and centered in a fixedsize image where the center of gravity of the intensity lies at the center of the image with 28 # 28 pixels. Thus, the dimensionality of each image sample vector is 28 * 28 = 784, where each element is binary. This is a relatively simple database for people who want to try machine learning techniques and pattern recognition methods on real-world data while spending minimal efforts on preprocessing and formatting. Using the references provided on the Web site, students and e ducators of machine learning can also benefit from a rather comprehensive set of machine learning literature with performance comparison readily available.
EVALUATION OF MACHINE LEARNING ALGORITHMS USING MNIST
General evaluation results on MNIST:
http://yann.lecun.com/exdb/mnist/
Details of logistic regression evaluated on MNIST:
http: //deeplearning.net/tutorial/logreg.html
Many well-known machine learning algorithms have been run on the MNIST database, so it is easy to assess the relative performance of a novel algorithm. The Web sitehttp://yann.lecun.com/exdb/mnist/ was updated in December of 2011 to list all major classification techniques and their results that were obtained using the MNIST database. In most experiments, the existing training data from the database were used in learning the classifiers, where ЁАnoneЁБ is entered in the ЁАPreprocessingЁБ column of the table on the Web site. In some experiments, the training set was augmented with artificially distorted versions of the original training samples. The distortions include random combinations of jittering, shifts, scaling, deskewing, deslanting, blurring, and compression. The type(s) of these and other distortions are specified in the ЁАPreprocessingЁБ column of the table as well.
A total of 68 classifiers are provided in the comparison table on the Web site, where ЁАTest Error Rates (%)ЁБ and links to the corresponding reference(s) are provided. These 68 machine learning techniques are organized into six broad categories:
Ёі linear classifiers
Ёі k-nearest neighbors
Ёі boosted stumps
Ёі nonlinear classifiers
Ёі support vector machines (SVMs)
Ёі neural nets (with no convolutional structure)
Ёі convolutional nets.
Each category contains up to 21 entries with very brief description of each in the ЁАClassifierЁБ column of the table. Much of the early techniques published in [1] are listed in the table.
BRIEF ANALYSIS OF THE MACHINE LEARNING ALGORITHMS EVALUATED ON MNIST
Comparing all the 68 classifiers listed on the MNIST Web site, we can make a brief analysis on the effectiveness of various techniques and of the preprocessing methods. Neural net classifiers tend to perform significantly better than other types of classifiers. Specifically, convolution structure in neural nets accounts for excellent classification performance. In fact, the record performance, about 0.27% error rate or 27 errors in the full 10,000 test set, is achieved by a committee of convolutional nets (with elastic distortion in augmenting the training set) [2]. Without the ЁАcommittee,ЁБ one single very large and deep convolutional neural net gives also a very low error rate of 0.35% [3]. The use of distortions, especially elastic distortion [4], to augment the training data is important to achieve very low error rates. Without such distortion, the error rate of a single large convolutional neural net is increased from 0.35% to 0.53% [5].
The depth of neural nets also accounts for low error rates. With both convolution structure and distortions, the deep versus shallow nets give the error rates of 0.35% [3] and 0.40ЈC0.60% [4],respectively.Without convolution structure and distortions or other types of special preprocessing, the lowest error rate in the literature, 0.83%, is achieved using the deep stacking/convex neural net [6]. The error rate is increased to 1.10% [7] when a corresponding shallow net is used.
Behind neural net techniques, k-nearest neighbor methods also produced low error rates, followed by virtual SVMs. Note that preprocessing is needed in both cases for the success
SUMMARY
The MNIST database gives a relatively simple static classification task for researchers and students to explore machine learning and pattern recognition techniques, saving unnecessary efforts on data preprocessing and formatting. This is analogous to the TIMIT database (a speech database created by Texas Instruments and Massachusetts Institute of Technology) familiar to most speech processing researchers in the signal processing community.
Just like the TIMIT phone classification and recognition tasks that have been productively used as a test bed for developing and testing speech recognition algorithms [7], MNIST has been used in a similar way for image and more general classification tasks. The Web site we introduce in this column provides the most comprehensive collection of resources for MNIST. In addition, this ЁАBest of the WebЁБ column also provides an analysis on a wide range of effective machine learning techniques evaluated on the MNIST task.
REFERENCES
[1] Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner,ЁАGradient-based learning applied to document recognition,ЁБ Proc. IEEE, vol. 86, no. 11, pp. 2278ЈC2324, Nov.1998.
[2] D. Ciresan, U. Meier, L. M. Gambardella, and J.Schmidhuber, ЁАConvolutional neural network committees for handwritten character classification,ЁБ in Proc. ICDAR, 2011.
[3] D. Ciresan, U. Meier, J. Masci, L. M. Gambardella, and J. Schmidhuber, ЁАFlexible, high performance convolutional neural networks for image classification,ЁБ in Proc. IJCAI, 2011.
[4] P. Simard, D. Steinkraus, and J. Platt, ЁАBest practices for convolutional neural networks applied to visual document analysis,ЁБ in Proc. ICDAR, 2003.
[5] K. Jarrett, K. Kavukcuoglu, M. Ranzato, and Y. LeCun, ЁА What is the best multi-stage architecture for object recognition?ЁБ in Proc. IEEE Int. Conf. Computer Vision (ICCV 2009).
[6] L. Deng and D. Yu, ЁАDeep convex network: A scalable architecture for speech pattern classification,ЁБ in Proc. Interspeech, Aug. 2011.
[7] L. Deng, D. Yu, and J. Platt, ЁАScalable stacking and learning for building deep architectures,ЁБ in Proc. ICASSP, Mar. 2012.
毕业66资料站 biye66.com ©2015-2026 版权所有 | 微信:15573586651 QQ:3903700237
本站毕业设计和毕业论文资料均属原创者所有,仅供学习交流之用,请勿转载并做其他非法用途.如有侵犯您的版权有损您的利益,请联系我们会立即改正或删除有关内容!