Artificial Intelligence
Article

How Claude Shannon Helped Kick-start Machine Learning

by
Industry Reports
February 9, 2022
Get Unlimited Access

You are puchasing a Digital Access Subscription. You will be automatically charged $9.99 every 28 days for one year. Your subscription will continue until you cancel. By subscribing, you are accepting the Automation Alley Terms of Service, Privacy Policy, and Terms of Sale.

SubscribeBecome a Member

Summary

A look at the long history of AI and digital transformation

Among the great engineers of the 20th century, who contributed the most to our 21st-century technologies? I say: Claude Shannon.

Shannon is best known for establishing the field of information theory. In a 1948 paper, one of the greatest in the history of engineering, he came up with a way of measuring the information content of a signal and calculating the maximum rate at which information could be reliably transmitted over any sort of communication channel. The article, titled “A Mathematical Theory of Communication,” describes the basis for all modern communications, including the wireless Internet on your smartphone and even an analog voice signal on a twisted-pair telephone landline. In 1966, the IEEE gave him its highest award, the Medal of Honor, for that work.

If information theory had been Shannon’s only accomplishment, it would have been enough to secure his place in the pantheon. But he did a lot more.

A decade before, while working on his master’s thesis at MIT, he invented the logic gate. At the time, electromagnetic relays—small devices that use magnetism to open and close electrical switches—were used to build circuits that routed telephone calls or controlled complex machines. However, there was no consistent theory on how to design or analyze such circuits. The way people thought about them was in terms of the relay coils being energized or not. Shannon showed that Boolean algebra could be used to move away from the relays themselves, into a more abstract understanding of the function of a circuit. He used this algebra of logic to analyze, and then synthesize, switching circuits and to prove that the overall circuit worked as desired. In his thesis he invented the AND, OR, and NOT logic gates. Logic gates are the building blocks of all digital circuits, upon which the entire edifice of computer science is based.

Read More Here

Industry Reports
Industry Reports

This denotes that a piece of content has been aggregated from an outside website.

Related
Become a Member