

This is a difficult project but very satisfying as it improves my understanding of what reasoning is and makes me amazed by our flexible and complicated thinking capabilities. We minimize the information to the what is relevant for us humans. That is, we look for the important factors that human expert would identify in order to find a solution equal to the one the computer found. We use a subtle model of human thinking to model how a human would make a recommendation based on the vast amount of data the computer searches. Image 1: DecodeChess explaining why the next best move for black is 23…Qg3
#Decode chess crack
Chess and AI have always been a hot topic, and we have been able to crack one of the most daunting riddles ever since the first AI-driven chess programs have been programmed: Why did the move the AI computer program (whose computation power far exceeds that of humans) decided to make is good and not another? As a co-founder at DecodeChess, my team develops algorithms that explain chess-playing program recommendations in a human way. I would like to share an example in which the focus on modelling human reasoning has resulted in an impressive XAI feat. So I’m here to make up for the neglect of science.” Oddly, those questions have been abandoned by science. When asked about the meaning of his new book’s title The Book of Why he answered: “ It means to be a summary of the work I’ve been doing the past 25 years about cause and effect, what it means in one’s life, its applications, and how we go about coming up with answers to questions that are inherently causal. Here again, Pearl accurately describes one instance of this neglect. This is the obligation and difficulty that we do not address nowadays. We no longer put enough efforts trying to solve problems in the area of general intelligence, or trying to create algorithms that actually mimic human way of reasoning. This brings us to the more profound problem – we abandoned an important goal of AI which is to understand human thinking. The last few years’ breakthroughs are the result of the mechanical ability of computers and not an ability to mimic high cognitive functions.

What is sometime missed in the (justified) enthusiasm from ML results is that people do it much better, or at least could have done it better if had they been equipped with the ability to contain enormous amounts of data. Rather, they are just useful statistical correlations.Ībsurdly, in the quest to achieve smart results we gave up intelligence. These mistakes are a strong hint that the features “learned” by these algorithms do not represent meaningful human casual connections between actual entities in the world. This is best seen in their humorous identification mistakes, for example identifying a toothbrush and a basketball as similar objects. Moreover, the patterns “learned” by NN statistical learning are a more useful tool for statistical correlations than a truly useful abstract.

This is possible, as humans can think and abstract about the data and be much more efficient than the current NN algorithms. This is an amazing ability of human thinking. We humans use just a few examples in order to get results that are almost as good as the algorithms. Actually, as Turing Award winner Judea Pearl correctly claims, they are more curve fitting techniques that are very useful but intellectually stupid: In order to perform reasonably, these algorithms need to digest an enormous amount of data. We think that they must be smart as they’re doing some really amazing things and they seem to represent the neural connection in our brains. With all this fuss around the amazing capabilities of the NN algorithms, we might think of NN and ML algorithms as mysterious and smart. Ironically, the more successful the algorithms, the more imminent is the problem of validating and explaining their results to a human user.Īs a result, we’re witnessing the rise of explainable AI (XAI), with companies and agencies putting more resources and budgets to explain these algorithms and give decision makers more control over AI. ML and NN algorithms’ lack of transparency, fairness, and safety (AKA the black box), together with their weaknesses in accounting for specific situations impedes the desired adoption rate of AI algorithms and creates frustration in different domains of our modern life. In the past two years, AI is suffering from an increasing problem of trust.
