IMTx: Understanding Artificial Intelligence through Algorithmic Information Theory
Can we characterize intelligent behavior?
Are there theoretical foundations on which Artificial Intelligence can be grounded?
About this course
Artificial Intelligence is more than just a collection of brilliant, innovative methods to solve problems.
If you are interested in machine learning or are planning to explore it, the course will make you see artificial learning in an entirely new way. You will know how to formulate optimal hypotheses for a learning task. And you will be able to analyze learning techniques such as clustering or neural networks as just ways of compressing information.
If you are interested in reasoning , you will understand that reasoning by analogy, reasoning by induction, explaining, proving, etc. are all alike; they all amount to providing more compact descriptions of situations.
If you are interested in mathematics , you will be amazed at the fact that crucial notions such as probability and randomness can be redefined in terms of algorithmic information. You will also understand that there are theoretical limits to what artificial intelligence can do.
If you are interested in human intelligence , you will find some intriguing results in this course. Thanks to algorithmic information, notions such as unexpectedness, interest and, to a certain extent, aesthetics, can be formally defined and computed, and this may change your views on what artificial intelligence can achieve in the future.
Half a century ago, three mathematicians made the same discovery independently. They understood that the concept of information belonged to computer science; that computer science could say what information means. Algorithmic Information Theory was born.
Algorithmic Information is what is left when all redundancy has been removed. This makes sense, as redundant content cannot add any useful information. Removing redundancy to extract meaningful information is something computer scientists are good at doing.
Algorithmic information is a great conceptual tool. It describes what artificial intelligence actually does , and what it should do to make optimal choices. It also says what artificial intelligence can’t do. Algorithmic information is an essential component in the theoretical foundations of AI.
Keywords:
Algorithmic information, Kolmogorov complexity, theoretical computer science, universal Turing machine, coding, compression, semantic distance, Zipf’s law, probability theory, algorithmic probability, computability, incomputability, random sequences, incompleteness theorem, machine learning, Occam’s razor, minimum description length, induction, cognitive science, relevance.
At a Glance:
Institution: IMTx
Subject: Computer Science
Level: Advanced
Prerequisites:
Examples of what you should know before embarking on this course
what a convex curve looks like,
that log(7^n) is n times log(7)
that rational numbers have finite or periodic expansion,
that rational numbers are countable, but that real numbers are not,
that the probability of “A and B” is the probability of “A knowing B” times the probability of B,
that 65 is 1000001 is in base 2 and 41 in base 16,
how to compute the sum of a finite geometric series,
that {‘a’:1, ‘i’:0} is a Python dictionary and why list(‘ab’*4)[::2] yields [‘a’,’a’,’a’,’a’],
that k-means is a clustering method,
what Bayes’ theorem tells us,
how Shannon’s information is related to probability,
that what is called a Turing machine is NOT the machine that Alan Turing (Benedict Cumberbatch)
is using in the movie The imitation game.
Language: English
Video Transcript: English
Associated skills:Probability, Basic Math, Information Theory, Cognitive Science, Probability Theories, Turing Machine, Planning, Artificial Intelligence, Machine Learning, Aesthetics, Innovation, Computer Science, Artificial Neural Networks
What You’ll Learn:
About this course
Artificial Intelligence is more than just a collection of brilliant, innovative methods to solve problems.
If you are interested in machine learning or are planning to explore it, the course will make you see artificial learning in an entirely new way. You will know how to formulate optimal hypotheses for a learning task. And you will be able to analyze learning techniques such as clustering or neural networks as just ways of compressing information.
If you are interested in reasoning , you will understand that reasoning by analogy, reasoning by induction, explaining, proving, etc. are all alike; they all amount to providing more compact descriptions of situations.
If you are interested in mathematics , you will be amazed at the fact that crucial notions such as probability and randomness can be redefined in terms of algorithmic information. You will also understand that there are theoretical limits to what artificial intelligence can do.
If you are interested in human intelligence , you will find some intriguing results in this course. Thanks to algorithmic information, notions such as unexpectedness, interest and, to a certain extent, aesthetics, can be formally defined and computed, and this may change your views on what artificial intelligence can achieve in the future.
Half a century ago, three mathematicians made the same discovery independently. They understood that the concept of information belonged to computer science; that computer science could say what information means. Algorithmic Information Theory was born.
Algorithmic Information is what is left when all redundancy has been removed. This makes sense, as redundant content cannot add any useful information. Removing redundancy to extract meaningful information is something computer scientists are good at doing.
Algorithmic information is a great conceptual tool. It describes what artificial intelligence actually does , and what it should do to make optimal choices. It also says what artificial intelligence can’t do. Algorithmic information is an essential component in the theoretical foundations of AI.
Keywords:
Algorithmic information, Kolmogorov complexity, theoretical computer science, universal Turing machine, coding, compression, semantic distance, Zipf’s law, probability theory, algorithmic probability, computability, incomputability, random sequences, incompleteness theorem, machine learning, Occam’s razor, minimum description length, induction, cognitive science, relevance.
There are no reviews yet.