Mark I Perceptron
This article has multiple issues. Please help improve it or discuss these issues on the talk page. (Learn how and when to remove these messages)
|
The Mark I Perceptron, developed by Frank Rosenblatt in 1958, was a pioneering supervised image classification learning system. It was the first implementation of an Artificial Intelligence (AI) machine. It differs from the Perceptron which is a software architecture proposed in 1943 by Warren McCulloch and Walter Pitts,[1] which was also employed in Mark I, and enhancements of which have continued to be an integral part of cutting edge AI technologies like the Transformer.
Architecture
[edit]The Mark I Perceptron was organized into three layers:[2]
- A set of sensory units which receive optical input
- A set of association units, each of which fire based on input from multiple sensory units
- A set of response units, which fire based on input from multiple association units
The connection between sensory units and association units were random. The working of association units was very similar to the response units.[2] Different versions of the Mark I used different numbers of units in each of the layers.[3]
Capabilities
[edit]In his 1957 proposal for development of the "Cornell Photoperceptron", Rosenblatt claimed:[4]
"Devices of this sort are expected ultimately to be capable of concept formation, language translation, collation of military intelligence, and the solution of problems through inductive logic."
With the first version of the Mark I Perceptron as early as 1958, Rosenblatt demonstrated a simple binary classification experiment, namely distinguishing between sheets of paper marked on the right versus those marked on the left side.[5]
One of the later experiments distinguished a square from a circle printed on paper. The shapes were perfect and their sizes fixed; the only variation was in their position and orientation. The Mark I Perceptron achieved 99.8% accuracy on a test dataset with 500 neurons in a single layer. The size of the training dataset was 10,000 example images. It took 3 seconds for the training pipeline to go through a single image. Higher accuracy was observed with thick outline figures compared to solid figures, likely because outline figures reduced overfitting.[3]
Another experiment distinguished between a square and a diamond for which 100% accuracy was achieved with only 60 training images, with a Perceptron having 1,000 neurons in a single layer. The time taken to process each training input for this larger perceptron was 15 seconds. The only variation was in position of the image, since rotation would have been ambiguous.
In that same experiment, it could distinguish between the letters X and E with 100% accuracy when trained with only 20 images (10 images of each letter). Variations in the images included both position and rotation by up to 30 degrees. When variation in rotation was increased to any angle (both in training and test datasets), the accuracy reduced to 90% with 60 training images (30 images of each letter).[3]
For distinguishing between the letters E and F, a more challenging problem due to their similarity, the same 1,000 neuron perceptron achieved an accuracy of more than 80% with 60 training images. Variation was only in the position of the image, with no rotation.[3]
References
[edit]- ^ McCulloch, Warren S.; Pitts, Walter (1943-12-01). "A logical calculus of the ideas immanent in nervous activity". The Bulletin of Mathematical Biophysics. 5 (4): 115–133. doi:10.1007/BF02478259. ISSN 1522-9602.
- ^ a b Rosenblatt, F. (1958). "The perceptron: A probabilistic model for information storage and organization in the brain". Psychological Review. 65 (6): 386–408. doi:10.1037/h0042519. ISSN 1939-1471. PMID 13602029.
- ^ a b c d Rosenblatt, Frank (March 1960). "Perceptron Simulation Experiments". Proceedings of the IRE. 48 (3): 301–309. doi:10.1109/JRPROC.1960.287598. ISSN 0096-8390.
- ^ Rosenblatt, Frank (January 1957). "The Perceptron—a perceiving and recognizing automaton" (PDF). Cornell Aeronautical Laboratory.
- ^ "Professor's perceptron paved the way for AI – 60 years too soon | Cornell Chronicle". news.cornell.edu. Retrieved 2024-10-08.
This article has not been added to any content categories. Please help out by adding categories to it so that it can be listed with similar articles. (November 2024) |