The History of Computing

hi thanks for tuning into singularity prosperity this video is the first in a multi-part series discussing computing and this video will be discussing the evolution of computing more specifically the evolution of the technologies that have brought upon the modern computing era the purpose of this video is so we can appreciate how fast technology is evolving and the people have brought us to this point many inventions have taken several centuries to develop into their modern forms and modern inventions are rarely the product of a single inventors efforts the computer is no different the bits and pieces of the computer both hardware and software have come together over many centuries with many people and groups each adding a small contribution we start as early as 3000 BC with the Chinese abacus how is this related to computing you ask the abacus was one of the first machines humans had ever created to be used for counting and calculating fast forward of 1642 and the abacus evolves into the first mechanical adding machine built by mathematician and scientist Blaise Pascal this first mechanical calculator the Pascaline is also where we see the first signs of technophobia emerging with mathematicians fearing the loss of their jobs due to progress also in the 1600s from the 1660s to the early 1700s we made Gottfried Leibniz a pioneer in many fields most notably known for his contributions to mathematics and considered by many the first computer scientists inspired by pesco he created his own calculating machine able to perform all four arithmetic operations he was also the first to lay down the concepts of binary arithmetic how all technology nowadays communicates and even envisioned a machine that used binary arithmetic from birth we are taught how to do arithmetic in base 10 and for most people that's all they're concerned with the numbers 0 to 9 however they're an infinite number of ways to represent information such as octal as base 8 hexadecimal as base 16 used represent colors base 256 which is used for encoding the list can go on binary is based to represented by the numbers 0 & 1 we'll explore later in this video why binary is essential for modern computing back on topic progressing to the 1800s were met with Charles Babbage Babbage is known as the father of the computer with the design of his mechanical calculating engines in 1820 Babbage notice that many computations consisted of operations that were regularly repeated and theorized that these operations can be done this led to his first design the difference engine it would have a fixed instruction set a fully automatic through the use of steam power and print its results into a table and 1830 Babbage stopped work on his Difference Engine to pursue a second idea the analytical engine elaborating on a difference engine this machine would be able to execute operations in non-numeric orders through the addition of conditional control store memory and read instructions from punch cards essentially making it a programmable mechanical computer unfortunately due to lack of funding his designs never came to reality but if they had would have sped up the invention of the computer by nearly 100 years also worth mentioning is a de Lovelace who worked very closely with Babbage she is considered the world's first programmer and came up with an algorithm that would calculate Bernoulli numbers and was designed to work with Babbage's machine she also outlined many fundamentals of programming such as data analysis looping and memory addressing 10 years prior to the turn of the century with inspiration from Babbage American inventor Herman Hollerith designed one of the first successful electromechanical machines referred to as a census tabulator this machine would read u.s. census data from punch cards up to 65 at a time and tally up the results Hollerith tabulator became so successful he wanted to found his own firm to market the device this company eventually became IBM to briefly explain how punch cards work essentially when fed into a machine an electrical connection is attempted to be made depending on where the holes in the card are will determine your input based on what connections are completed to input data to the punch card you could use a key punch machine aka the first iteration of a keyboard the eighteen hundred's are a period where the theory of computing began to evolve and machines started to be used for calculations but the 1900s is where we begin to see the pieces of this nearly 5,000 year puzzle coming together especially between 1930 to 1950 in 1936 Alan Turing proposed a concept of a universal machine later to be dubbed the Turing machine capable of computing anything that is computable up to this point machines are not only able to do certain tasks to the hardware was designed for the concept of the modern computer is largely based off turns ideas also starting in 1936 German engineer Konrad Zuse invented the world's first programmable computer this device read instructions from punched tape and was the first computer to use boolean logic and binary to make decisions through the use of relays for reference boolean logic is simply logic that results either a true or false output or one corresponding to binary one or zero we'll be diving into boolean logic deeper later in this video Zeus will later use punch cards to encode information binary essentially making them the first data storage and memory devices in 1942 with the computer the z4 Zeus also released the world's first commercial computer for these reasons many consider Zeus the inventor of the modern-day computer in 1937 Howard Aiken was colleagues at Harvard and collaboration with IBM began work on the Harvard mark 1 calculating machine a programmable calculator and inspired by Babbage's analytical engine this machine was composed of nearly 1 million parts had over 500 miles of wiring and weighed nearly 5 tons the mark 1 had 60 sets of 24 switches for manual data entry and could store 72 numbers each 23 decimal digits long I could do 3 additions or subtractions in a second a multiplication took 6 seconds a division took 15 point three seconds and a logarithm or trig function took about one minute as a funny side note one of the primary programmers of the Mark 1 Grace Hopper discovered the first computer bug a dead moth blocking one of the reading holes of the machine hopper is also credited with coining the word debugging the vacuum tube error marks the beginning of modern computing the first technology that was fully digital and unlike the relays used in previous computers or less power-hungry faster and more reliable beginning in 1937 and completing in 1942 the first digital computer was built by John Atanasoff and his graduate student Clifford berry the computer was dubbed the ABC unlike previously built computers like those built by Zeus the ABC was purely digital it used vacuum tubes and included binary math and boolean logic to solve up to 29 equations at a time in 1943 the Colossus was built in collaboration with Alan Turing to assist in breaking German crypto codes not to be confused of Turing's bombe that actually solved the Enigma this computer was fully digital as well but unlike the ABC was fully programmable making it the first fully programmable digital computer completing construction in 1946 the electrical numerical integrator and computer aka the ENIAC was completed composed of nearly 18,000 vacuum tubes and large enough to fill an entire room the antiochus consider the first successful high-speed electronic digital computer it was somewhat program but like Aikens Marquand was a painter a wire every time the instruction set had to be changed the ENIAC essentially took the concepts from Atanasoff ABC and elaborated on them in a much larger scale meanwhile the ENIAC was under construction in 1945 mathematician John von Neumann contributed a new understanding of how computers should be organized and built further elaborating on Turing stories and bringing clarity to the idea from Peter memory and addressing he elaborated on conditional addressing or subroutines something Babbage had envisioned for his analytical engine nearly 100 years earlier also the idea that instructions or the program running on a computer could be modified in the same way as data and decode them in binary von Neumann assisted in the design of the ENIAC successor the electronic discrete variable automatic computer aka the edvac which was completed in 1950 and the first stored-program computer it was able to operate over 1,000 instructions per second he is also credited with being the father of computer viral adji was his design of a self reproducing computer program and it contains essentially those things which the modern computer has in it although in somewhat primitive form this machine has stored program concept as its major feature and that in fact is the thing which makes the modern computer revolution possible at this point you can see that computing had officially evolved into its own field from mechanical to electromechanical relays I took milliseconds to digital vacuum tubes that took only microseconds from binary as a way to encode information with punch cards to being used with boolean logic and represented by physical technologies like relays and vacuum tubes to finally being used to store instructions and programs from the abacus as a way to count to Pascal's mechanical calculator the theories of Lebanese Alan Turing and John von Neumann division of Babbage and the intellect of Lovelace George Bulls contribution of boolean logic the progressing inventions of a programmable calculator to a stored-program fully digital computer and countless other inventions individuals and groups each step of further accumulation of knowledge while the title of the inventor of the computer may be given to an individual or group it was really a joint contribution over 5,000 years and more so between 1802 1950 vacuum tubes are a huge improvement over relays but they still didn't make economic sense in a large scale for example of the NEX 18000 tubes roughly 50 would burn out per day and around-the-clock team of technicians would be needed to replace them vacuum tubes are also the reason why computers took up the space of entire rooms weighed multiple tons and consumed enough energy to power a small town in 1947 the first silicon transistor was invented at Bell Labs and by 1954 the first transistorized digital computer was invented aka the truddick it was composed of 800 transistors took the space of point zero at five cubic meters compared to the 28 the ENIAC took up only took 100 watts of power and could reform 1 million operations per second also during this era we begin to see major introductions on both the hardware and software aspect of computing on the hardware side the first memory device the random-access magnetic horse store was introduced in 1951 by Jay Forrester in other words the beginnings of what is now known as Ram today the first hard drive was introduced by IBM in 1957 it weighed one ton and could store five megabytes costing approximately 27,000 dollars per month in today's money on the software side is where a lot of major innovations and breakthroughs began to come this because computer hardware and architecture was beginning to become more standardized instead of everyone working on different variations of a computing machine assembly was a first programming language to be introduced in 1949 I really started taking off in this era of computing assembly was a way to communicate with the machine in pseudo English instead of machine language aka binary the first true widely used programming language was Fortran invented by John Bacchus at IBM in 1954 assembly as a low-level language and Fortran is a high-level language and low-level languages while you aren't writing instructions in machine code a very deep understanding of computer architecture and instructions is still required to execute a desired program which means a limited number of people have the skills and it is very error-prone also in the early to mid 50s to compile code back to machine code was still an expensive and time-consuming process this all changed with a Grace Hopper and a development of the first computer compiler hopper if you remember from earlier also found the first computer bug that's a lot for programming of repeaters to become more affordable and instantaneous instead of the time-consuming process of writing code in assembly and then manually converting it back to machine code as a side note Hopper assisted with the invention of and other early programming language cobalt this era marks the beginnings of the modern computing era and where the exponential trend of computing performance really began while transistors were a major improvement over vacuum tubes they still have to be individually soldered together as a result the more complex computers became left a more complicated and numerous connections between transistors increasing the likelihood of faulty wiring and 1958 this all changed with Jack Kilby of Texas Instruments and his invention of the integrated circuit the integrated circuit was a way to pack many transistors onto a single chip instead of individually wiring transistors packing all the transistors also significantly reduced the power and heat consumption of computers once again and made them significantly more economically feasible to design and buy integrated circuits sparked a hardware revolution and beyond Computers assisted in the development of various other electronic devices due to miniaturization such as a mouse invented by Douglas Engelbart in 1964 he also demonstrated the first graphical user interface as a sidenote computer speed performance memory and storage also began to iteratively increases ICS could pack more transistors into smaller surface areas that's demonstrated by the invention of the floppy disk in 1971 by IBM and in the same year DRAM by Intel – list of you along with hardware further advances in software were made as well with an explosion of programming languages and the introduction of some of the most common languages today basic in 1964 and C in 1971 as you can see from throughout this video computing since the 1900s has evolved at an increasingly fast rate thus in 1965 led Gordon Moore one of the founders of Intel to make one of the greatest predictions in human history computing power would double every two months at low cost and that computers would eventually be so small that that could be embedded into homes cars and what he referred to as personal portable communications equipment aka mobile phones we now refer to this as Moore's law Harrison charges to further illustrate how fast computing was evolving and what more based is predictions on one of my colleagues called this Moore's law rather than just being something that chronicles the progress of the industry it kind of became something that drove the progress of the industry a tremendous amount of engineering and commitment has been required to make that happen but much to my surprise the industry has been able to keep up with the projection at this point the video has come down conclusion I'd like to thank you for taking the time to watch it if you enjoyed it please leave a thumbs up and if you want me to elaborate on any of the topics discussed or have any topic suggestions please leave them in the comments below consider subscribing to my channel for more content follow my medium publication for accompanying blogs and like my facebook page for more bite-sized chunks of content this has been encore you've been watching singularity prosperity and I'll see you again soon

22 thoughts on “The History of Computing

  1. Join our Discord server for much better community discussions!

  2. Why did you leave out the Jaquard loom? The jaquard loom, which used punch cards, inspired Babbage to make his machines.

  3. hey what the fuck – you didnt mention Robert Noyce and integrated circuits! what the fuck?! Fuck jack-off kilby!

  4. keep it up, the youngsters of today need to know where there processing power started that is in their devices today!

  5. if you ever go back in time to 1950. don't tell them you got 128 Gb smartphone in your pocket, you will be locked as crazy person

  6. 3:57 "capable of computing anything that is computable" noooo seriously i how wonderful and extraordinary is that !! It's like I'm capable of eating anythinggg that is eatable 🤣 .. no but really nice video i just laughed a lot at that point

Leave a Reply

Your email address will not be published. Required fields are marked *