The Internet of Things (IoT), is a phrase that was first coined in the late 1990s. It is the art of making a seemingly “stupid” device “smart” and thus giving it the ability to transfer data over a network from machine to machine.One of the first of these “smart” devices was a Coca-Cola vending machine in 1982. As the story goes, a group of thirsty, frustrated, and innovative computer science students at Carnegie Mellon University in Pittsburgh, Pennsylvania wanted to know the inventory level of their nearby Coca-Cola machine. They installed micro switches that relayed the inventory levels on the machine’s racks and also reported whether the newly loaded drinks were cold yet. The concept of IoT first became popular outside computer science circles in 1999 due to the Auto-ID Centre at the Massachusetts Institute of Technology, as well as similar markets, publishing analyses on the subject. Beginning with radio-frequency identification, the applications soon spread to technologies such as barcodes, near field communication, digital watermarking and QR codes. Today some common applications can be found in the consumer world and would include examples such as wearable technology that reports to your tablet or smart phone, while tracking an exercise plan. Or what about hive by British Gas that allows you to control your heating from your smart device. The concept of IoT and its application, however, isn’t as new as it might sound. Parcel delivery companies adopted this idea decades ago. It’s why you can track your package from its origination point along every stop of its journey until it reaches your doorstep. Read more about the Internet of Things:
- Government wants firms and cities to be inventive with £10m Internet of Things fund
- Is the Internet of Things the new entrepreneurial battleground?
- Why the Generation of Things is more valuable to the economy than Internet of Things
Share this story