In artificial intelligence (AI), commonsense reasoning is a human-like ability to make presumptions about the type and essence of ordinary situations humans encounter every day. These assumptions include judgments about the nature of physical objects, taxonomic properties, and peoples' intentions. A device that exhibits commonsense reasoning might be capable of drawing conclusions that are similar to humans' folk psychology (humans' innate ability to reason about people's behavior and intentions) and naive physics (humans' natural understanding of the physical world).
Some definitions and characterizations of common sense from different authors include:
"Commonsense knowledge includes the basic facts about events (including actions) and their effects, facts about knowledge and how it is obtained, facts about beliefs and desires. It also includes the basic facts about material objects and their properties."
"Commonsense knowledge differs from encyclopedic knowledge in that it deals with general knowledge rather than the details of specific entities."
Commonsense knowledge is "real world knowledge that can provide a basis for additional knowledge to be gathered and interpreted automatically".
The commonsense world consists of "time, space, physical interactions, people, and so on".
Common sense is "all the knowledge about the world that we take for granted but rarely state out loud".
Common sense is "broadly reusable background knowledge that's not specific to a particular subject area... knowledge that you ought to have."
NYU professor Ernest Davis characterizes commonsense knowledge as "what a typical seven year old knows about the world", including physical objects, substances, plants, animals, and human society. It usually excludes book-learning, specialized knowledge, and knowledge of conventions; but it sometimes includes knowledge about those topics. For example, knowing how to play cards is specialized knowledge, not "commonsense knowledge"; but knowing that people play cards for fun does count as "commonsense knowledge".
This page is automatically generated and may contain information that is not correct, complete, up-to-date, or relevant to your search query. The same applies to every other page on this website. Please make sure to verify the information with EPFL's official sources.
Natural language processing is ubiquitous in modern intelligent technologies, serving as a foundation for language translators, virtual assistants, search engines, and many more. In this course, stude
This course will develop logical reasoning and argumentation skills to enable you to influence decision making. You will achieve this by learning how to represent and communicate your reasoning as ar
We introduce formal verification as an approach for developing highly reliable systems. Formal verification finds proofs that computer systems work under all relevant scenarios. We will learn how to u
Explores natural language generation, focusing on building systems that produce coherent text for human consumption using various decoding methods and evaluation metrics.
The Turing test, originally called the imitation game by Alan Turing in 1950, is a test of a machine's ability to exhibit intelligent behaviour equivalent to, or indistinguishable from, that of a human. Turing proposed that a human evaluator would judge natural language conversations between a human and a machine designed to generate human-like responses. The evaluator would be aware that one of the two partners in conversation was a machine, and all participants would be separated from one another.
Deep learning is part of a broader family of machine learning methods, which is based on artificial neural networks with representation learning. The adjective "deep" in deep learning refers to the use of multiple layers in the network. Methods used can be either supervised, semi-supervised or unsupervised.
In computer science, in particular in knowledge representation and reasoning and metalogic, the area of automated reasoning is dedicated to understanding different aspects of reasoning. The study of automated reasoning helps produce computer programs that allow computers to reason completely, or nearly completely, automatically. Although automated reasoning is considered a sub-field of artificial intelligence, it also has connections with theoretical computer science and philosophy.
A critical component of a successful language generation pipeline is the decoding algorithm. However, the general principles that should guide the choice of a decoding algorithm re- main unclear. Previous works only compare decoding algorithms in narrow sc ...
2023
, , , , , ,
Language models (LMs) have recently shown remarkable performance on reasoning tasks by explicitly generating intermediate inferences, e.g., chain-of-thought prompting. However, these intermediate inference steps may be in- appropriate deductions from the i ...
2023
,
The recent advance of large language models (LLMs) demonstrates that these large-scale foundation models achieve remarkable capabilities across a wide range of language tasks and domains. The success of the statistical learning approach challenges our unde ...