mirror of
https://github.com/MindWorkAI/AI-Studio.git
synced 2025-02-05 21:49:06 +00:00
5935 lines
329 KiB
C#
5935 lines
329 KiB
C#
namespace AIStudio.Settings.DataModel;
|
||
|
||
/// <summary>
|
||
/// Catalog of biases.
|
||
/// </summary>
|
||
/// <remarks>
|
||
/// Based on the work of Buster Benson, John Manoogian III, and Brian
|
||
/// Rene Morrissette. The biases were clustered and organized by
|
||
/// Buster Benson. The texts originally come from Wikipedia and
|
||
/// were reduced to a short definition by Brian Rene Morrissette.
|
||
/// John Manoogian III designed the original poster from Buster
|
||
/// Benson's work, which was then supplemented with definitions
|
||
/// by Brian Rene Morrissette.
|
||
///
|
||
/// All texts were checked by Thorsten Sommer against the 2024
|
||
/// version of Wikipedia. Most texts were replaced with the latest
|
||
/// versions, and long texts were shortened to the essentials.
|
||
///
|
||
/// The texts were revised and, when necessary, supplemented by
|
||
/// Thorsten Sommer for integration into AI Studio. Sources and
|
||
/// additional links were also researched by Thorsten Sommer.
|
||
/// </remarks>
|
||
public static class BiasCatalog
|
||
{
|
||
public static readonly Bias NONE = new()
|
||
{
|
||
Id = Guid.Empty,
|
||
Category = BiasCategory.NONE,
|
||
Name = "None",
|
||
Description = "No bias selected.",
|
||
Related = [],
|
||
Links = [],
|
||
};
|
||
|
||
#region WHAT_SHOULD_WE_REMEMBER
|
||
|
||
private static readonly Bias MISATTRIBUTION_OF_MEMORY = new()
|
||
{
|
||
Id = new Guid("dd45c762-0599-4c6d-82e0-d10f7ee85bb1"),
|
||
Category = BiasCategory.WHAT_SHOULD_WE_REMEMBER,
|
||
Name = "Misattribution of Memory",
|
||
Description =
|
||
"""
|
||
# Misattribution of Memory
|
||
The ability to remember information correctly, but being wrong about the source of
|
||
that information. Includes the following three sub-effects:
|
||
|
||
- Source Confusion:
|
||
Source confusion is an attribute seen in different people’s accounts of the same
|
||
event after hearing people speak about the situation. An example of this would
|
||
be a witness who heard a police officer say he had a gun and then that witness
|
||
later says they saw the gun even though they didn’t. The source of the memory
|
||
is the police officer’s testimony, not actual perception.
|
||
|
||
- Cryptomnesia:
|
||
Individuals mistakenly believe that they are the original generators of the
|
||
thought.
|
||
|
||
- False Memory:
|
||
False memories occur when a person’s identity and interpersonal relationships
|
||
are strongly centered around a memory of an experience that did not actually
|
||
take place. False memories are often the result of leading questions in a
|
||
therapeutic practice termed Recovered Memory Therapy. In this practice,
|
||
psychiatrists often put their patients under hypnosis to recover repressed
|
||
memories. This can be detrimental, as the individual may recall memories that
|
||
never occurred.
|
||
""",
|
||
|
||
Related = [],
|
||
Links = [
|
||
"https://en.wikipedia.org/wiki/Misattribution_of_memory",
|
||
"https://en.wikipedia.org/wiki/Cryptomnesia",
|
||
"https://en.wikipedia.org/wiki/Source-monitoring_error",
|
||
"https://en.wikipedia.org/wiki/False_memory",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias LIST_LENGTH_EFFECT = new()
|
||
{
|
||
Id = new Guid("688bba31-0b8e-49c5-8693-aecb37018a08"),
|
||
Category = BiasCategory.WHAT_SHOULD_WE_REMEMBER,
|
||
Name = "List Length Effect",
|
||
Description =
|
||
"""
|
||
# List Length Effect
|
||
The finding that recognition performance for a short list is superior to that for
|
||
a long list.
|
||
""",
|
||
|
||
Related = [],
|
||
Links = [
|
||
"https://en.wikipedia.org/wiki/List_of_cognitive_biases#Other_memory_biases",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias MISINFORMATION_EFFECT = new()
|
||
{
|
||
Id = new Guid("2b69b071-6587-4ea1-a4f5-aee4e2fef43c"),
|
||
Category = BiasCategory.WHAT_SHOULD_WE_REMEMBER,
|
||
Name = "Misinformation Effect",
|
||
Description =
|
||
"""
|
||
# Misinformation Effect
|
||
When a person's recall of episodic memories becomes less accurate because of
|
||
post-event information. The misinformation effect is an example of retroactive
|
||
interference which occurs when information presented later interferes with the
|
||
ability to retain previously encoded information. Individuals have also been
|
||
shown to be susceptible to incorporating misleading information into their memory
|
||
when it is presented within a question. Essentially, the new information that a
|
||
person receives works backward in time to distort memory of the original event.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("dd45c762-0599-4c6d-82e0-d10f7ee85bb1"), // MISATTRIBUTION_OF_MEMORY -> False Memory
|
||
new Guid("4d377bac-062a-46d3-a1a5-46f3ac804a97"), // SUGGESTIBILITY
|
||
],
|
||
Links = [
|
||
"https://en.wikipedia.org/wiki/Misinformation_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias LEVELING_AND_SHARPENING = new()
|
||
{
|
||
Id = new Guid("d1ed47f9-2415-4fa3-8ca3-151e9e4ee3ca"),
|
||
Category = BiasCategory.WHAT_SHOULD_WE_REMEMBER,
|
||
Name = "Leveling and Sharpening",
|
||
Description =
|
||
"""
|
||
# Leveling and Sharpening
|
||
Leveling occurs when you hear or remember something, and drop details which do not fit cognitive
|
||
categories and/or assumptions; sharpening occurs when you hear or remember something, and emphasize
|
||
details which do fit cognitive categories and/or assumptions.
|
||
""",
|
||
|
||
Related = [],
|
||
Links = [
|
||
"https://en.wikipedia.org/wiki/Leveling_and_sharpening",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias PEAK_END_RULE = new()
|
||
{
|
||
Id = new Guid("cf71d1e1-f49e-4d8f-a6c3-37056297bf13"),
|
||
Category = BiasCategory.WHAT_SHOULD_WE_REMEMBER,
|
||
Name = "Peak-End Rule",
|
||
Description =
|
||
"""
|
||
# Peak-End Rule
|
||
The peak–end rule is a psychological heuristic in which people judge an experience largely based on how
|
||
they felt at its peak (i.e., its most intense point) and at its end, rather than based on the total sum
|
||
or average of every moment of the experience. The effect occurs regardless of whether the experience is
|
||
pleasant or unpleasant. To the heuristic, other information aside from that of the peak and end of the
|
||
experience is not lost, but it is not used. This includes net pleasantness or unpleasantness and how
|
||
long the experience lasted. The peak–end rule is thereby a specific form of the more general extension
|
||
neglect and duration neglect.
|
||
""",
|
||
|
||
Related = [],
|
||
Links = [
|
||
"https://en.wikipedia.org/wiki/Peak%E2%80%93end_rule",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias FADING_AFFECT_BIAS = new()
|
||
{
|
||
Id = new Guid("0378a05c-b55b-4451-a7f4-b5e1d6287d83"),
|
||
Category = BiasCategory.WHAT_SHOULD_WE_REMEMBER,
|
||
Name = "Fading Affect Bias",
|
||
Description =
|
||
"""
|
||
# Fading Affect Bias
|
||
The fading affect bias, more commonly known as FAB, is a psychological phenomenon in which memories
|
||
associated with negative emotions tend to be forgotten more quickly than those associated with positive
|
||
emotions. FAB only refers to the feelings one has associated with the memories and not the content of
|
||
the memories themselves.
|
||
""",
|
||
|
||
Related = [],
|
||
Links = [
|
||
"https://en.wikipedia.org/wiki/Fading_affect_bias",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias NEGATIVITY_BIAS = new()
|
||
{
|
||
Id = new Guid("ef521fbb-c20b-47c9-87f8-a571a06a03eb"),
|
||
Category = BiasCategory.WHAT_SHOULD_WE_REMEMBER,
|
||
Name = "Negativity Bias",
|
||
Description =
|
||
"""
|
||
# Negativity Bias
|
||
The negativity bias, also known as the negativity effect, is a cognitive bias that, even when positive
|
||
or neutral things of equal intensity occur, things of a more negative nature (e.g. unpleasant thoughts,
|
||
emotions, or social interactions; harmful/traumatic events) have a greater effect on one's psychological
|
||
state and processes than neutral or positive things. In other words, something very positive will
|
||
generally have less of an impact on a person's behavior and cognition than something equally emotional but
|
||
negative. The negativity bias has been investigated within many different domains, including the formation
|
||
of impressions and general evaluations; attention, learning, and memory; and decision-making and risk
|
||
considerations.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("ad3ed908-c56e-411b-a130-8af8574ff67b"), // LOSS_AVERSION
|
||
],
|
||
Links = [
|
||
"https://en.wikipedia.org/wiki/Negativity_bias",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias PREJUDICE = new()
|
||
{
|
||
Id = new Guid("efb6606f-4629-4e5e-973f-94d5ac496638"),
|
||
Category = BiasCategory.WHAT_SHOULD_WE_REMEMBER,
|
||
Name = "Prejudice",
|
||
Description =
|
||
"""
|
||
# Prejudice
|
||
Prejudice can be an affective feeling towards a person based on their perceived group membership. The word
|
||
is often used to refer to a preconceived (usually unfavourable) evaluation or classification of another
|
||
person based on that person's perceived personal characteristics, such as political affiliation, sex,
|
||
gender, gender identity, beliefs, values, social class, age, disability, religion, sexuality, race,
|
||
ethnicity, language, nationality, culture, complexion, beauty, height, body weight, occupation, wealth,
|
||
education, criminality, sport-team affiliation, music tastes or other perceived characteristics.
|
||
|
||
The word "prejudice" can also refer to unfounded or pigeonholed beliefs and it may apply to
|
||
"any unreasonable attitude that is unusually resistant to rational influence". Gordon Allport defined
|
||
prejudice as a "feeling, favorable or unfavorable, toward a person or thing, prior to, or not based on,
|
||
actual experience". Auestad defines prejudice as characterized by "symbolic transfer", transfer of
|
||
a value-laden meaning content onto a socially-formed category and then on to individuals who are taken to
|
||
belong to that category, resistance to change, and overgeneralization.
|
||
|
||
The United Nations Institute on Globalization, Culture and Mobility has highlighted research considering
|
||
prejudice as a global security threat due to its use in scapegoating some populations and inciting others
|
||
to commit violent acts towards them and how this can endanger individuals, countries, and the international
|
||
community.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Prejudice",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias IMPLICIT_STEREOTYPES = new()
|
||
{
|
||
Id = new Guid("30bd6403-b7f4-4d16-9494-af6a22b349d3"),
|
||
Category = BiasCategory.WHAT_SHOULD_WE_REMEMBER,
|
||
Name = "Implicit Stereotypes",
|
||
Description =
|
||
"""
|
||
# Implicit Stereotypes
|
||
The unconscious attribution of particular qualities to a member of a certain social group. Implicit stereotypes are
|
||
influenced by experience, and are based on learned associations between various qualities and social categories,
|
||
including race or gender. Individuals' perceptions and behaviors can be affected by implicit stereotypes, even
|
||
without the individuals' intention or awareness.
|
||
|
||
An implicit bias or implicit stereotype is the pre-reflective attribution of particular qualities by an individual
|
||
to a member of some social out group. Implicit stereotypes are thought to be shaped by experience and based on learned
|
||
associations between particular qualities and social categories, including race and/or gender. Individuals' perceptions
|
||
and behaviors can be influenced by the implicit stereotypes they hold, even if they are sometimes unaware they hold such
|
||
stereotypes. Implicit bias is an aspect of implicit social cognition: the phenomenon that perceptions, attitudes, and
|
||
stereotypes can operate prior to conscious intention or endorsement. The existence of implicit bias is supported by a
|
||
variety of scientific articles in psychological literature. Implicit stereotype was first defined by psychologists
|
||
Mahzarin Banaji and Anthony Greenwald in 1995.
|
||
|
||
Explicit stereotypes, by contrast, are consciously endorsed, intentional, and sometimes controllable thoughts and beliefs.
|
||
|
||
Implicit biases, however, are thought to be the product of associations learned through past experiences. Implicit
|
||
biases can be activated by the environment and operate prior to a person's intentional, conscious endorsement. Implicit
|
||
bias can persist even when an individual rejects the bias explicitly.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Implicit_stereotype",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias IMPLICIT_ASSOCIATIONS = new()
|
||
{
|
||
Id = new Guid("6f1d8a61-cb64-44fe-9a52-5ee66c22fba4"),
|
||
Category = BiasCategory.WHAT_SHOULD_WE_REMEMBER,
|
||
Name = "Implicit Associations",
|
||
Description =
|
||
"""
|
||
# Implicit Associations
|
||
A person's automatic association between mental representations of objects (concepts) in memory.
|
||
*Controversial. This is not a bias, it is an association algorithm.*
|
||
|
||
Related: The implicit-association test
|
||
The implicit-association test (IAT) is an assessment intended to detect subconscious associations
|
||
between mental representations of objects (concepts) in memory. Its best-known application is the
|
||
assessment of implicit stereotypes held by test subjects, such as associations between particular
|
||
racial categories and stereotypes about those groups. The test has been applied to a variety of
|
||
belief associations, such as those involving racial groups, gender, sexuality, age, and religion
|
||
but also the self-esteem, political views, and predictions of the test taker.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Implicit_stereotype#Activation_of_implicit_stereotypes",
|
||
"https://en.wikipedia.org/wiki/Implicit-association_test",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias SPACING_EFFECT = new()
|
||
{
|
||
Id = new Guid("41e06aaf-73c2-4f48-9962-312d57308468"),
|
||
Category = BiasCategory.WHAT_SHOULD_WE_REMEMBER,
|
||
Name = "Spacing Effect",
|
||
Description =
|
||
"""
|
||
# Spacing Effect
|
||
The spacing effect demonstrates that learning is more effective when study sessions are spaced out. This
|
||
effect shows that more information is encoded into long-term memory by spaced study sessions, also known
|
||
as spaced repetition or spaced presentation, than by massed presentation ("cramming").
|
||
|
||
The phenomenon was first identified by Hermann Ebbinghaus, and his detailed study of it was published in
|
||
the 1885 book "Über das Gedächtnis. Untersuchungen zur experimentellen Psychologie" (Memory: A Contribution
|
||
to Experimental Psychology), which suggests that active recall with increasing time intervals reduces the
|
||
probability of forgetting information. This robust finding has been supported by studies of many explicit
|
||
memory tasks such as free recall, recognition, cued-recall, and frequency estimation.
|
||
|
||
Researchers have offered several possible explanations of the spacing effect, and much research has been
|
||
conducted that supports its impact on recall. In spite of these findings, the robustness of this phenomenon
|
||
and its resistance to experimental manipulation have made empirical testing of its parameters difficult.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Spacing_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias SUGGESTIBILITY = new()
|
||
{
|
||
Id = new Guid("4d377bac-062a-46d3-a1a5-46f3ac804a97"),
|
||
Category = BiasCategory.WHAT_SHOULD_WE_REMEMBER,
|
||
Name = "Suggestibility",
|
||
Description =
|
||
"""
|
||
# Suggestibility
|
||
The quality of being inclined to accept and act on the suggestions of others where false but plausible
|
||
information is given and one fills in the gaps in certain memories with false information when recalling
|
||
a scenario or moment. When the subject has been persistently told something about a past event, his or
|
||
her memory of the event conforms to the repeated message.
|
||
|
||
Suggestibility can be seen in people's day-to-day lives:
|
||
|
||
- Someone witnesses an argument after school. When later asked about the "huge fight" that occurred, he
|
||
recalls the memory, but unknowingly distorts it with exaggerated fabrications, because he now thinks
|
||
of the event as a "huge fight" instead of a simple argument.
|
||
|
||
- Children are told by their parents they are good singers, so from then on they believe they are talented
|
||
while their parents were in fact being falsely encouraging.
|
||
|
||
- A teacher could trick his AP Psychology students by saying, "Suggestibility is the distortion of memory
|
||
through suggestion or misinformation, right?" It is likely that the majority of the class would agree
|
||
with him because he is a teacher and what he said sounds correct. However, the term is really the
|
||
definition of the misinformation effect.
|
||
|
||
However, suggestibility can also be seen in extremes, resulting in negative consequences:
|
||
|
||
- A witness' testimony is altered because the police or attorneys make suggestions during the interview,
|
||
which causes their already uncertain observations to become distorted memories.
|
||
|
||
- A young girl begins suffering migraines which lead to sleep deprivation and depression. Her therapist,
|
||
a specialist in cases involving child abuse, repeatedly asks her whether her father had sexually abused
|
||
her. This suggestion causes the young girl to fabricate memories of her father molesting her, which lead
|
||
to her being placed in foster care and her father being tried on charges of abuse.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("2b69b071-6587-4ea1-a4f5-aee4e2fef43c"), // MISINFORMATION_EFFECT,
|
||
new Guid("dd45c762-0599-4c6d-82e0-d10f7ee85bb1"), // MISATTRIBUTION_OF_MEMORY -> False Memory
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Suggestibility",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias TIP_OF_THE_TONGUE_PHENOMENON = new()
|
||
{
|
||
Id = new Guid("ad341a56-ffa5-4dd1-b3c6-ef2476b89b5a"),
|
||
Category = BiasCategory.WHAT_SHOULD_WE_REMEMBER,
|
||
Name = "Tip of the Tongue Phenomenon",
|
||
Description =
|
||
"""
|
||
# Tip of the Tongue Phenomenon
|
||
Tip of the tongue (also known as TOT, or lethologica) is the phenomenon of failing to retrieve a word or term
|
||
from memory, combined with partial recall and the feeling that retrieval is imminent. The phenomenon's name
|
||
comes from the saying, "It's on the tip of my tongue." The tip of the tongue phenomenon reveals that
|
||
lexical access occurs in stages.
|
||
|
||
People experiencing the tip-of-the-tongue phenomenon can often recall one or more features of the target word,
|
||
such as the first letter, its syllabic stress, and words similar in sound, meaning, or both sound and meaning.
|
||
Individuals report a feeling of being seized by the state, feeling something like mild anguish while
|
||
searching for the word, and a sense of relief when the word is found. While many aspects of the
|
||
tip-of-the-tongue state remain unclear, there are two major competing explanations for its occurrence:
|
||
the direct-access view and the inferential view. Emotion and the strength of the emotional ties to what
|
||
is trying to be remembered can also have an impact on the TOT phenomenon. The stronger the emotional ties,
|
||
the longer it takes to retrieve the item from memory.
|
||
|
||
TOT states should be distinguished from FOK (feeling of knowing) states. FOK, in contrast, is the feeling
|
||
that one will be able to recognize — from a list of items — an item that is currently inaccessible. There
|
||
are still currently opposing hypotheses in the psychological literature regarding the separability of the
|
||
process underlying these concepts. However, there is some evidence that TOTs and FOKs draw on different
|
||
parts of the brain.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Tip_of_the_tongue",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias GOOGLE_EFFECT = new()
|
||
{
|
||
Id = new Guid("aca65269-40cc-4a9f-b850-c0e2eb283987"),
|
||
Category = BiasCategory.WHAT_SHOULD_WE_REMEMBER,
|
||
Name = "Google Effect",
|
||
Description =
|
||
"""
|
||
# Google Effect
|
||
The Google effect, also called digital amnesia, is the tendency to forget information that can be found
|
||
readily online by using Internet search engines. According to the first study about the Google effect, people
|
||
are less likely to remember certain details they believe will be accessible online. However, the study also
|
||
claims that people's ability to learn information offline remains the same. This effect may also be seen
|
||
as a change to what information and what level of detail is considered to be important to remember.
|
||
|
||
In a big replication study published in Nature 2018, the Google effect was one of the experiments which
|
||
could not be replicated.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Google_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias NEXT_IN_LINE_EFFECT = new()
|
||
{
|
||
Id = new Guid("c0b3d9f9-c0d9-482f-bf6f-d52dffe58205"),
|
||
Category = BiasCategory.WHAT_SHOULD_WE_REMEMBER,
|
||
Name = "Next-In-Line Effect",
|
||
Description =
|
||
"""
|
||
# Next-In-Line Effect
|
||
The next-in-line effect is the phenomenon of people being unable to recall information concerning events
|
||
immediately preceding their turn to perform. The effect was first studied experimentally by Malcolm Brenner
|
||
in 1973. In his experiment the participants were each in turn reading a word aloud from an index card, and
|
||
after 25 words were asked to recall as many of all the read words as possible. The results of the experiment
|
||
showed that words read aloud within approximately nine seconds before the subject's own turn were recalled
|
||
worse than other words.
|
||
|
||
The reason for the next-in-line effect appears to be a deficit in encoding the perceived information preceding
|
||
a performance. That is, the information is never stored to long-term memory and thus cannot be retrieved later
|
||
after the performance. One finding supporting this theory is that asking the subjects beforehand to pay more
|
||
attention to events preceding their turn to perform can prevent the memory deficit and even result in
|
||
overcompensation, making people remember the events before their turn better than others.
|
||
|
||
In addition, the appearance of the next-in-line effect does not seem to be connected to the level of fear of
|
||
negative evaluation. Both people with lower and higher anxiety levels are subject to the memory deficit.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Next-in-line_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias TESTING_EFFECT = new()
|
||
{
|
||
Id = new Guid("2fe5fbe7-3fff-4621-9111-21d8c3b8bcb2"),
|
||
Category = BiasCategory.WHAT_SHOULD_WE_REMEMBER,
|
||
Name = "Testing Effect",
|
||
Description =
|
||
"""
|
||
# Testing Effect
|
||
The testing effect (also known as retrieval practice, active recall, practice testing, or test-enhanced learning)
|
||
suggests long-term memory is increased when part of the learning period is devoted to retrieving information from
|
||
memory. It is different from the more general practice effect, defined in the APA Dictionary of Psychology as
|
||
"any change or improvement that results from practice or repetition of task items or activities."
|
||
|
||
Cognitive psychologists are working with educators to look at how to take advantage of tests—not as an assessment
|
||
tool, but as a teaching tool since testing prior knowledge is more beneficial for learning when compared to only
|
||
reading or passively studying material (even more so when the test is more challenging for memory).
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Testing_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias ABSENT_MINDEDNESS = new()
|
||
{
|
||
Id = new Guid("f459e613-4f6a-472e-ab9a-0961f5f4a685"),
|
||
Category = BiasCategory.WHAT_SHOULD_WE_REMEMBER,
|
||
Name = "Absent-Mindedness",
|
||
Description =
|
||
"""
|
||
# Absent-Mindedness
|
||
In the field of psychology, absent-mindedness is a mental state wherein a person is forgetfully inattentive.
|
||
It is the opposite mental state of mindfulness. Absentmindedness is often caused by things such as boredom,
|
||
sleepiness, rumination, distraction, or preoccupation with one's own internal monologue. When experiencing
|
||
absent-mindedness, people exhibit signs of memory lapses and weak recollection of recent events.
|
||
|
||
Absent-mindedness can usually be a result of a variety of other conditions often diagnosed by clinicians,
|
||
such as attention deficit hyperactivity disorder and depression. In addition to absent-mindedness leading
|
||
to an array of consequences affecting daily life, it can have more severe, long-term problems.
|
||
|
||
It can have three different causes:
|
||
1) a low level of attention ("blanking" or "zoning out");
|
||
2) intense attention to a single object of focus (hyperfocus) that makes a person oblivious to events around him or her;
|
||
3) unwarranted distraction of attention from the object of focus by irrelevant thoughts or environmental events.
|
||
|
||
Absent-mindedness is also noticed as a common characteristic of personalities with schizoid personality
|
||
disorder.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Absent-mindedness",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias LEVELS_OF_PROCESSING_EFFECT = new()
|
||
{
|
||
Id = new Guid("a4027640-1f52-4ff1-ae13-bd14a30d5b8d"),
|
||
Category = BiasCategory.WHAT_SHOULD_WE_REMEMBER,
|
||
Name = "Levels of Processing Effect",
|
||
Description =
|
||
"""
|
||
# Levels of Processing Effect
|
||
The Levels of Processing model, created by Fergus I. M. Craik and Robert S. Lockhart in 1972, describes memory recall
|
||
of stimuli as a function of the depth of mental processing. More analysis produce more elaborate and stronger memory
|
||
than lower levels of processing. Depth of processing falls on a shallow to deep continuum. Shallow processing (e.g.,
|
||
processing based on phonemic and orthographic components) leads to a fragile memory trace that is susceptible to rapid
|
||
decay. Conversely, deep processing (e.g., semantic processing) results in a more durable memory trace. There are three
|
||
levels of processing in this model. (1) Structural processing, or visual, is when we remember only the physical quality
|
||
of the word (e.g. how the word is spelled and how letters look). (2) Phonemic processing includes remembering the word
|
||
by the way it sounds (e.g. the word tall rhymes with fall). (3) Lastly, we have semantic processing in which we encode
|
||
the meaning of the word with another word that is similar or has similar meaning. Once the word is perceived, the
|
||
brain allows for a deeper processing.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("4f61b9fa-146a-4b6e-b075-f0ba2ee0d9d0"), // PROCESSING_DIFFICULTY_EFFECT
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Levels_of_Processing_model",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias SUFFIX_EFFECT = new()
|
||
{
|
||
Id = new Guid("1f2b459b-26bc-4a2d-b48e-a9b06d34f924"),
|
||
Category = BiasCategory.WHAT_SHOULD_WE_REMEMBER,
|
||
Name = "Suffix Effect",
|
||
Description =
|
||
"""
|
||
# Suffix Effect
|
||
The selective impairment in recall of the final items of a spoken list when the list is followed by a
|
||
nominally irrelevant speech item, or suffix.
|
||
|
||
Diminishment of the recency effect because a sound item is appended to the list that the subject is not
|
||
required to recall. A form of serial position effect.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/List_of_cognitive_biases#Other_memory_biases",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias SERIAL_POSITION_EFFECT = new()
|
||
{
|
||
Id = new Guid("fdd1e260-125b-4d06-bab5-a6204f96d5a7"),
|
||
Category = BiasCategory.WHAT_SHOULD_WE_REMEMBER,
|
||
Name = "Serial-Position Effect",
|
||
Description =
|
||
"""
|
||
# Serial-Position Effect
|
||
Serial-position effect is the tendency of a person to recall the first and last items in a series best,
|
||
and the middle items worst. When asked to recall a list of items in any order (free recall), people tend
|
||
to begin recall with the end of the list, recalling those items best (the recency effect). Among earlier
|
||
list items, the first few items are recalled more frequently than the middle items (the primacy effect).
|
||
|
||
One suggested reason for the primacy effect is that the initial items presented are most effectively stored
|
||
in long-term memory because of the greater amount of processing devoted to them. (The first list item can
|
||
be rehearsed by itself; the second must be rehearsed along with the first, the third along with the first
|
||
and second, and so on.) The primacy effect is reduced when items are presented quickly and is enhanced
|
||
when presented slowly (factors that reduce and enhance processing of each item and thus permanent storage).
|
||
Longer presentation lists have been found to reduce the primacy effect.
|
||
|
||
One theorised reason for the recency effect is that these items are still present in working memory when
|
||
recall is solicited. Items that benefit from neither (the middle items) are recalled most poorly. An
|
||
additional explanation for the recency effect is related to temporal context: if tested immediately after
|
||
rehearsal, the current temporal context can serve as a retrieval cue, which would predict more recent items
|
||
to have a higher likelihood of recall than items that were studied in a different temporal context (earlier
|
||
in the list). The recency effect is reduced when an interfering task is given. Intervening tasks involve
|
||
working memory, as the distractor activity, if exceeding 15 to 30 seconds in duration, can cancel out the
|
||
recency effect. Additionally, if recall comes immediately after the test, the recency effect is consistent
|
||
regardless of the length of the studied list, or presentation rate.
|
||
|
||
Amnesiacs with poor ability to form permanent long-term memories do not show a primacy effect, but do show
|
||
a recency effect if recall comes immediately after study. People with Alzheimer's disease exhibit a
|
||
reduced primacy effect but do not produce a recency effect in recall.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("741cafef-3f47-45dd-a082-bb243eba124a"), // RECENCY_EFFECT,
|
||
new Guid("78f65dab-ab6d-4c4c-81f5-250edd1e8991"), // PRIMACY_EFFECT
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Serial-position_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias PART_LIST_CUING = new()
|
||
{
|
||
Id = new Guid("005b650d-74be-4c10-a279-33dcd5c13f84"),
|
||
Category = BiasCategory.WHAT_SHOULD_WE_REMEMBER,
|
||
Name = "Part-List Cuing",
|
||
Description =
|
||
"""
|
||
# Part-List Cuing
|
||
The re-exposure of a subset of learned material as a retrieval cue can impair recall of the remaining material.
|
||
""",
|
||
|
||
Related = [],
|
||
Links = [],
|
||
};
|
||
|
||
private static readonly Bias RECENCY_EFFECT = new()
|
||
{
|
||
Id = new Guid("741cafef-3f47-45dd-a082-bb243eba124a"),
|
||
Category = BiasCategory.WHAT_SHOULD_WE_REMEMBER,
|
||
Name = "Recency Effect",
|
||
Description =
|
||
"""
|
||
# Recency Bias
|
||
Recency bias is a cognitive bias that favors recent events over historic ones; a memory bias. Recency bias
|
||
gives "greater importance to the most recent event", such as the final lawyer's closing argument a jury hears
|
||
before being dismissed to deliberate.
|
||
|
||
Recency bias should not be confused with anchoring or confirmation bias. Recency bias is related to the
|
||
serial-position effect known as the recency effect. It is not to be confused with recency illusion, the
|
||
belief or impression that a word or language usage is of recent origin when in reality it is long-established.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("fdd1e260-125b-4d06-bab5-a6204f96d5a7"), // SERIAL_POSITION_EFFECT
|
||
new Guid("78f65dab-ab6d-4c4c-81f5-250edd1e8991"), // PRIMACY_EFFECT
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Recency_bias",
|
||
"https://en.wikipedia.org/wiki/Serial-position_effect#Recency_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias PRIMACY_EFFECT = new()
|
||
{
|
||
Id = new Guid("78f65dab-ab6d-4c4c-81f5-250edd1e8991"),
|
||
Category = BiasCategory.WHAT_SHOULD_WE_REMEMBER,
|
||
Name = "Primacy Effect",
|
||
Description =
|
||
"""
|
||
# Primacy Effect
|
||
In psychology and sociology, the primacy effect (also known as the primacy bias) is a cognitive bias that results
|
||
in a subject recalling primary information presented better than information presented later on. For example, a
|
||
subject who reads a sufficiently long list of words is more likely to remember words toward the beginning than
|
||
words in the middle.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("fdd1e260-125b-4d06-bab5-a6204f96d5a7"), // SERIAL_POSITION_EFFECT
|
||
new Guid("741cafef-3f47-45dd-a082-bb243eba124a"), // RECENCY_EFFECT
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Serial-position_effect#Primacy_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias MEMORY_INHIBITION = new()
|
||
{
|
||
Id = new Guid("0a370e78-860b-4784-9acf-688b5e1c3148"),
|
||
Category = BiasCategory.WHAT_SHOULD_WE_REMEMBER,
|
||
Name = "Memory Inhibition",
|
||
Description =
|
||
"""
|
||
# Memory Inhibition
|
||
In psychology, memory inhibition is the ability not to remember irrelevant information. The scientific concept of memory
|
||
inhibition should not be confused with everyday uses of the word "inhibition". Scientifically speaking, memory inhibition
|
||
is a type of cognitive inhibition, which is the stopping or overriding of a mental process, in whole or in part, with or
|
||
without intention.
|
||
|
||
Memory inhibition is a critical component of an effective memory system. While some memories are retained for a lifetime,
|
||
most memories are forgotten. According to evolutionary psychologists, forgetting is adaptive because it facilitates
|
||
selectivity of rapid, efficient recollection. For example, a person trying to remember where they parked their car
|
||
would not want to remember every place they have ever parked. In order to remember something, therefore, it is essential
|
||
not only to activate the relevant information, but also to inhibit irrelevant information.
|
||
|
||
*Controversial. This is not a bias, it is a logical information sorting algorithm.*
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("4e571eaf-7c2b-44c8-b8cb-0c8da658b82d"), // FREQUENCY_ILLUSION
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Memory_inhibition",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias MODALITY_EFFECT = new()
|
||
{
|
||
Id = new Guid("eeca14c3-8710-4522-8991-81db170d7f8b"),
|
||
Category = BiasCategory.WHAT_SHOULD_WE_REMEMBER,
|
||
Name = "Modality Effect",
|
||
Description =
|
||
"""
|
||
# Modality Effect
|
||
The modality effect is a term used in experimental psychology, most often in the fields dealing with memory and learning,
|
||
to refer to how learner performance depends on the presentation mode of studied items. Modality can refer to a number of
|
||
characteristics of the presented study material. However, this term is usually used to describe the improved recall of
|
||
the final items of a list when that list is presented verbally in comparison with a visual representation.
|
||
|
||
Some studies use the term modality to refer to a general difference in performance based upon the mode of presentation.
|
||
For example, Gibbons demonstrated modality effects in an experiment by making participants count either beeping sounds
|
||
or visually presented dots. In his book about teaching Mathematics, Craig Barton refers to the Modality Effect, arguing
|
||
that students learn better when images or narrations are presented alongside verbal narration, as opposed to being
|
||
presented with on screen text.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Modality_effect",
|
||
],
|
||
};
|
||
|
||
#endregion
|
||
|
||
#region TOO_MUCH_INFORMATION
|
||
|
||
private static readonly Bias AVAILABILITY_HEURISTIC = new()
|
||
{
|
||
Id = new Guid("d749ce96-32f3-4c3d-86f7-26ff4edabe4a"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Availability Heuristic",
|
||
Description =
|
||
"""
|
||
# Availability Heuristic
|
||
A mental shortcut that relies on immediate examples that come to a given person’s mind when evaluating
|
||
a specific topic, concept, method or decision. The availability heuristic operates on the notion that
|
||
if something can be recalled, it must be important, or at least more important than alternative
|
||
solutions which are not as readily recalled. Subsequently, under the availability heuristic, people
|
||
tend to heavily weigh their judgments toward more recent information, making new opinions biased
|
||
toward that latest news.
|
||
|
||
The mental availability of an action's consequences is positively related to those consequences'
|
||
perceived magnitude. In other words, the easier it is to recall the consequences of something,
|
||
the greater those consequences are often perceived to be. Most notably, people often rely on the
|
||
content of their recall if its implications are not called into question by the difficulty they
|
||
have in recalling it.
|
||
|
||
After seeing news stories about child abductions, people may judge that the likelihood of this event
|
||
is greater. Media coverage can help fuel a person's example bias with widespread and extensive coverage
|
||
of unusual events, such as homicide or airline accidents, and less coverage of more routine, less
|
||
sensational events, such as common diseases or car accidents. For example, when asked to rate the
|
||
probability of a variety of causes of death, people tend to rate "newsworthy" events as more likely
|
||
because they can more readily recall an example from memory. Moreover, unusual and vivid events
|
||
like homicides, shark attacks, or lightning are more often reported in mass media than common and
|
||
un-sensational causes of death like common diseases.
|
||
|
||
For example, many people think that the likelihood of dying from shark attacks is greater than that
|
||
of dying from being hit by falling airplane parts when more people actually die from falling airplane
|
||
parts. When a shark attack occurs, the deaths are widely reported in the media whereas deaths as a
|
||
result of being hit by falling airplane parts are rarely reported in the media.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Availability_heuristic",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias ATTENTIONAL_BIAS = new()
|
||
{
|
||
Id = new Guid("368cc51b-a235-4fa4-ad90-446c084ffae8"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Attentional Bias",
|
||
Description =
|
||
"""
|
||
# Attentional Bias
|
||
Attentional bias refers to how a person's perception is affected by selective factors in their attention.
|
||
Attentional biases may explain an individual's failure to consider alternative possibilities when occupied
|
||
with an existing train of thought. For example, cigarette smokers have been shown to possess an attentional
|
||
bias for smoking-related cues around them, due to their brain's altered reward sensitivity. Attentional bias
|
||
has also been associated with clinically relevant symptoms such as anxiety and depression.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Attentional_bias",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias ILLUSORY_TRUTH_EFFECT = new()
|
||
{
|
||
Id = new Guid("cadafb8f-d1ed-4c92-9c29-2f1cb0797a66"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Illusory Truth Effect",
|
||
Description =
|
||
"""
|
||
# Illusory Truth Effect
|
||
The illusory truth effect (also known as the illusion of truth effect, validity effect, truth effect,
|
||
or the reiteration effect) is the tendency to believe false information to be correct after repeated
|
||
exposure. This phenomenon was first identified in a 1977 study at Villanova University and Temple
|
||
University. When truth is assessed, people rely on whether the information is in line with their
|
||
understanding or if it feels familiar. The first condition is logical, as people compare new information
|
||
with what they already know to be true. Repetition makes statements easier to process relative to new,
|
||
unrepeated statements, leading people to believe that the repeated conclusion is more truthful. The
|
||
illusory truth effect has also been linked to hindsight bias, in which the recollection of confidence
|
||
is skewed after the truth has been received.
|
||
|
||
In a 2015 study, researchers discovered that familiarity can overpower rationality and that repetitively
|
||
hearing that a certain statement is wrong can paradoxically cause it to feel right. Researchers
|
||
observed the illusory truth effect's impact even on participants who knew the correct answer to begin
|
||
with but were persuaded to believe otherwise through the repetition of a falsehood, to "processing fluency".
|
||
|
||
The illusory truth effect plays a significant role in fields such as advertising, news media, and
|
||
political propaganda.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("b9edd2f0-8503-4eb5-a4c3-369fcb318894"), // HINDSIGHT_BIAS
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Illusory_truth_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias MERE_EXPOSURE_EFFECT = new()
|
||
{
|
||
Id = new Guid("8b6cd991-fcf4-4e45-b3ac-f20987667d94"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Mere-Exposure Effect",
|
||
Description =
|
||
"""
|
||
# Mere-Exposure Effect
|
||
The mere-exposure effect is a psychological phenomenon by which people tend to develop liking or disliking for
|
||
things merely because they are familiar with them. In social psychology, this effect is sometimes called the
|
||
familiarity principle. The effect has been demonstrated with many kinds of things, including words, Chinese
|
||
characters, paintings, pictures of faces, geometric figures, and sounds. In studies of interpersonal
|
||
attraction, the more often people see a person, the more pleasing and likeable they find that person.
|
||
|
||
The most obvious application of the mere-exposure effect is in advertising, but research on its effectiveness at
|
||
enhancing consumer attitudes toward particular companies and products has been mixed.
|
||
|
||
The mere-exposure effect exists in most areas of human decision-making. For example, many stock traders tend
|
||
to invest in securities of domestic companies merely because they are more familiar with them, even though
|
||
international markets offer similar or better alternatives. The mere-exposure effect also distorts the
|
||
results of journal-ranking surveys; academics who previously published or completed reviews for a particular
|
||
academic journal rate it dramatically higher than those who did not. There are mixed results on the question
|
||
of whether mere exposure can promote good relations between different social groups. When groups already
|
||
have negative attitudes to each other, further exposure can increase hostility. A statistical analysis of
|
||
voting patterns found that candidates' exposure has a strong effect on the number of votes they receive,
|
||
distinct from the popularity of their policies.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Mere-exposure_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias CONTEXT_EFFECT = new()
|
||
{
|
||
Id = new Guid("ccba2bca-8739-4b05-8c88-e54424e441d4"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Context Effect",
|
||
Description =
|
||
"""
|
||
# Context Effect
|
||
A context effect is an aspect of cognitive psychology that describes the influence of environmental factors
|
||
on one's perception of a stimulus. The impact of context effects is considered to be part of top-down
|
||
design. The concept is supported by the theoretical approach to perception known as constructive perception.
|
||
Context effects can impact our daily lives in many ways such as word recognition, learning abilities, memory,
|
||
and object recognition. It can have an extensive effect on marketing and consumer decisions. For example,
|
||
research has shown that the comfort level of the floor that shoppers are standing on while reviewing products
|
||
can affect their assessments of product's quality, leading to higher assessments if the floor is comfortable
|
||
and lower ratings if it is uncomfortable. Because of effects such as this, context effects are currently
|
||
studied predominantly in marketing.
|
||
|
||
Context effects can have a wide range of impacts in daily life. In reading difficult handwriting context
|
||
effects are used to determine what letters make up a word. This helps us analyze potentially ambiguous
|
||
messages and decipher them correctly. It can also affect our perception of unknown sounds based on the noise
|
||
in the environment. For example, we may fill in a word we cannot make out in a sentence based on the
|
||
other words we could understand. Context can prime our attitudes and beliefs about certain topics based
|
||
on current environmental factors and our previous experiences with them.
|
||
|
||
Context effects also affect memory. We are often better able to recall information in the location in which
|
||
we learned it or studied it. For example, while studying for a test it is better to study in the environment
|
||
that the test will be taken in (i.e. classroom) than in a location where the information was not learned
|
||
and will not need to be recalled. This phenomenon is called transfer-appropriate processing.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Context_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias CUE_DEPENDENT_FORGETTING = new()
|
||
{
|
||
Id = new Guid("944bc142-895e-4c7f-ba00-bbceefc383c9"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Cue-Dependent Forgetting",
|
||
Description =
|
||
"""
|
||
# Cue-Dependent Forgetting
|
||
Cue-dependent forgetting, or retrieval failure, is the failure to recall information without memory cues.
|
||
The term either pertains to semantic cues, state-dependent cues or context-dependent cues.
|
||
|
||
Upon performing a search for files in a computer, its memory is scanned for words. Relevant files containing
|
||
this word or string of words are displayed. This is not how memory in the human mind works. Instead,
|
||
information stored in the memory is retrieved by way of association with other memories. Some memories
|
||
can not be recalled by simply thinking about them. Rather, one must think about something associated
|
||
with it.
|
||
|
||
For example, if someone tries and fails to recollect the memories they had about a vacation they went on,
|
||
and someone mentions the fact that they hired a classic car during this vacation, this may make them remember
|
||
all sorts of things from that trip, such as what they ate there, where they went and what books they read.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Cue-dependent_forgetting",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias STATE_DEPENDENT_MEMORY = new()
|
||
{
|
||
Id = new Guid("bf83101d-47af-4d81-8306-935d4ab59fd7"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "State-Dependent Memory",
|
||
Description =
|
||
"""
|
||
# State-Dependent Memory
|
||
State-dependent memory or state-dependent learning is the phenomenon where people remember more information
|
||
if their physical or mental state is the same at time of encoding and time of recall. State-dependent memory
|
||
is heavily researched in regards to its employment both in regards to synthetic states of consciousness
|
||
(such as under the effects of psychoactive drugs) as well as organic states of consciousness such as mood.
|
||
While state-dependent memory may seem rather similar to context-dependent memory, context-dependent memory
|
||
involves an individual's external environment and conditions (such as the room used for study and to take
|
||
the test) while state-dependent memory applies to the individual's internal conditions (such as use of
|
||
substances or mood).
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("6b049d68-9104-4579-a7a4-a744c11bd65f"), // CONTEXT_DEPENDENT_MEMORY
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/State-dependent_memory",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias CONTEXT_DEPENDENT_MEMORY = new()
|
||
{
|
||
Id = new Guid("6b049d68-9104-4579-a7a4-a744c11bd65f"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Context-Dependent Memory",
|
||
Description =
|
||
"""
|
||
# Context-Dependent Memory
|
||
In psychology, context-dependent memory is the improved recall of specific episodes or information when the
|
||
context present at encoding and retrieval are the same. In a simpler manner, "when events are represented
|
||
in memory, contextual information is stored along with memory targets; the context can therefore cue memories
|
||
containing that contextual information". One particularly common example of context-dependence at work occurs
|
||
when an individual has lost an item (e.g. lost car keys) in an unknown location. Typically, people try to
|
||
systematically "retrace their steps" to determine all of the possible places where the item might be located.
|
||
Based on the role that context plays in determining recall, it is not at all surprising that individuals often
|
||
quite easily discover the lost item upon returning to the correct context. This concept is heavily related to
|
||
the encoding specificity principle.
|
||
|
||
This example best describes the concept of context-dependent forgetting. However, the research literature on
|
||
context-dependent memory describes a number of different types of contextual information that may affect
|
||
recall such as environmental context-dependent memory, state-dependent learning, cognitive context-dependent
|
||
memory and mood-congruent memory. Research has also shown that context-dependence may play an important role
|
||
in numerous situations, such as memory for studied material, or events that have occurred following the
|
||
consumption of alcohol or other drugs.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("bf83101d-47af-4d81-8306-935d4ab59fd7"), // STATE_DEPENDENT_MEMORY
|
||
new Guid("ccba2bca-8739-4b05-8c88-e54424e441d4"), // CONTEXT_EFFECT
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Context-dependent_memory",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias FREQUENCY_ILLUSION = new()
|
||
{
|
||
Id = new Guid("4e571eaf-7c2b-44c8-b8cb-0c8da658b82d"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Frequency Illusion",
|
||
Description =
|
||
"""
|
||
# Frequency Illusion
|
||
The frequency illusion (also known as the Baader–Meinhof phenomenon) is a cognitive bias in which a
|
||
person notices a specific concept, word, or product more frequently after recently becoming aware of it.
|
||
|
||
The name "Baader–Meinhof phenomenon" was coined in 1994 by Terry Mullen in a letter to the St. Paul
|
||
Pioneer Press. The letter describes how, after mentioning the name of the German terrorist group
|
||
Baader–Meinhof once, he kept noticing it. This led to other readers sharing their own experiences of
|
||
the phenomenon, leading it to gain recognition. It was not until 2005, when Stanford linguistics
|
||
professor Arnold Zwicky wrote about this effect on his blog, that the name "frequency illusion"
|
||
was coined.
|
||
|
||
The main cause behind frequency illusion, and other related illusions and biases, seems to be
|
||
selective attention. Selective attention refers to the process of selecting and focusing on
|
||
selective objects while ignoring distractions. This means that people have the
|
||
unconscious cognitive ability to filter for what they are focusing on.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("af63ce77-f6c6-4e0f-8a9e-3daedc497f9a"), // CONFIRMATION_BIAS
|
||
new Guid("0a370e78-860b-4784-9acf-688b5e1c3148"), // MEMORY_INHIBITION
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Frequency_illusion",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias HOT_COLD_EMPATHY_GAP = new()
|
||
{
|
||
Id = new Guid("e4e091cf-fed3-4c09-9c21-509db0b2729b"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Hot-Cold Empathy Gap",
|
||
Description =
|
||
"""
|
||
# Hot-Cold Empathy Gap
|
||
A hot-cold empathy gap is a cognitive bias in which people underestimate the influences of visceral
|
||
drives on their own attitudes, preferences, and behaviors. It is a type of empathy gap. The most
|
||
important aspect of this idea is that human understanding is "state-dependent". For example, when
|
||
one is angry, it is difficult to understand what it is like for one to be calm, and vice versa;
|
||
when one is blindly in love with someone, it is difficult to understand what it is like for one
|
||
not to be, (or to imagine the possibility of not being blindly in love in the future). Importantly,
|
||
an inability to minimize one's gap in empathy can lead to negative outcomes in medical settings
|
||
(e.g., when a doctor needs to accurately diagnose the physical pain of a patient).
|
||
|
||
Hot-cold empathy gaps can be analyzed according to their direction:
|
||
|
||
- Hot-to-cold: People under the influence of visceral factors (hot state) do not fully grasp how
|
||
much their behavior and preferences are being driven by their current state; they think instead
|
||
that these short-term goals reflect their general and long-term preferences.
|
||
|
||
- Cold-to-hot: People in a cold state have difficulty picturing themselves in hot states, minimizing
|
||
the motivational strength of visceral impulses. This leads to unpreparedness when visceral forces
|
||
inevitably arise.
|
||
|
||
They can also be classified in regards to their relation with time (past or future) and whether they
|
||
occur intra- or inter-personally:
|
||
|
||
- intrapersonal prospective: the inability to effectively predict their own future behavior when in
|
||
a different state. See also projection bias.
|
||
|
||
- intrapersonal retrospective: when people recall or try to understand behaviors that happened in a
|
||
different state.
|
||
|
||
- interpersonal: the attempt to evaluate behaviors or preferences of another person who is in a
|
||
state different from one's own.
|
||
|
||
Visceral factors
|
||
Visceral factors are an array of influences which include hunger, thirst, love, sexual arousal,
|
||
drug cravings for the drugs one is addicted to, physical pain, and desire for revenge. These
|
||
drives have a disproportionate effect on decision making and behavior: the mind, when affected
|
||
(i.e., in a hot state), tends to ignore all other goals in an effort to placate these influences.
|
||
These states can lead a person to feel "out of control" and act impulsively.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("61ca5b76-66d0-4ce2-b260-7fd42696000a"), // PROJECTION_BIAS
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Hot-cold_empathy_gap",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias OMISSION_BIAS = new()
|
||
{
|
||
Id = new Guid("ad32d669-fc79-44c9-a570-609e1ccdc799"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Omission Bias",
|
||
Description =
|
||
"""
|
||
# Omission Bias
|
||
Omission bias is the phenomenon in which people prefer omission (inaction) over commission (action)
|
||
and people tend to judge harm as a result of commission more negatively than harm as a result of
|
||
omission. It can occur due to a number of processes, including psychological inertia, the perception
|
||
of transaction costs, and the perception that commissions are more causal than omissions. In social
|
||
political terms the Universal Declaration of Human Rights establishes how basic human rights are to
|
||
be assessed in article 2, as "without distinction of any kind, such as race, colour, sex, language,
|
||
religion, political or other opinion, national or social origin, property, birth or other status."
|
||
criteria that are often subject to one or another form of omission bias. It is controversial as to
|
||
whether omission bias is a cognitive bias or is often rational. The bias is often showcased through
|
||
the trolley problem and has also been described as an explanation for the endowment effect and
|
||
status quo bias.
|
||
|
||
A real-world example is when parents decide not to vaccinate their children because of the potential
|
||
chance of death—even when the probability the vaccination will cause death is much less likely than
|
||
death from the disease prevented.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("b81482f8-b2cf-4b86-a5a4-fcd29aee4e69"), // ENDOWMENT_EFFECT
|
||
new Guid("b9e05a25-ac09-407d-8aee-f54a04decf0b"), // STATUS_QUO_BIAS
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Omission_bias",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias BASE_RATE_FALLACY = new()
|
||
{
|
||
Id = new Guid("1de0de03-a2a7-4248-b004-4152d84a3c86"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Base Rate Fallacy",
|
||
Description =
|
||
"""
|
||
# Base Rate Fallacy
|
||
The base rate fallacy, also called base rate neglect or base rate bias, is a type of fallacy in
|
||
which people tend to ignore the base rate (e.g., general prevalence) in favor of the individuating
|
||
information (i.e., information pertaining only to a specific case). For example, if someone hears
|
||
that a friend is very shy and quiet, they might think the friend is more likely to be a librarian
|
||
than a salesperson, even though there are far more salespeople than librarians overall - hence
|
||
making it more likely that their friend is actually a salesperson. Base rate neglect is a specific
|
||
form of the more general extension neglect.
|
||
|
||
Another example: Students were asked to estimate the GPAs of hypothetical students. When given
|
||
relevant statistics about GPA distribution, students tended to ignore them if given descriptive
|
||
information about the particular student even if the new descriptive information was obviously
|
||
of little or no relevance to GPA.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("8533edf9-3117-48c5-8f78-efbd996911f0"), // CONSERVATISM
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Base_rate_fallacy",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias BIZARRENESS_EFFECT = new()
|
||
{
|
||
Id = new Guid("7f2d9bd2-96e5-4100-85f8-a13b37e91a9f"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Bizarreness Effect",
|
||
Description =
|
||
"""
|
||
# Bizarreness Effect
|
||
Bizarreness effect is the tendency of bizarre material to be better remembered than common material.
|
||
The scientific evidence for its existence is contested. Some research suggests it does exist, some
|
||
suggests it doesn't exist and some suggests it leads to worse remembering.
|
||
|
||
McDaniel and Einstein argues that bizarreness intrinsically does not enhance memory in their paper
|
||
from 1986. They claim that bizarre information becomes distinctive. It is the distinctiveness that
|
||
according to them makes encoding easier. Which makes common sense from an instinctual perspective
|
||
as the human brain will disregard ingesting information it already is familiar with and will be
|
||
particularly attuned to taking in new information as an adaptation technique.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Bizarreness_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias HUMOUR_EFFECT = new()
|
||
{
|
||
Id = new Guid("4c5ef2d4-5ebb-48ea-b9ee-9b2751ae6914"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Humour Effect",
|
||
Description =
|
||
"""
|
||
# Humour Effect
|
||
The tendency to better remember humorous items than non-humorous ones.
|
||
""",
|
||
|
||
Related = [],
|
||
Links = [],
|
||
};
|
||
|
||
private static readonly Bias VON_RESTORFF_EFFECT = new()
|
||
{
|
||
Id = new Guid("b922da6f-765e-42e9-b675-f8109c010f2f"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Von Restorff Effect",
|
||
Description =
|
||
"""
|
||
# Von Restorff Effect
|
||
The Von Restorff effect, also known as the "isolation effect", predicts that when multiple
|
||
homogeneous stimuli are presented, the stimulus that differs from the rest is more likely
|
||
to be remembered. The theory was coined by German psychiatrist and pediatrician Hedwig
|
||
von Restorff (1906–1962), who, in her 1933 study, found that when participants were presented
|
||
with a list of categorically similar items with one distinctive, isolated item on the list,
|
||
memory for the item was improved.
|
||
|
||
For example, if a person examines a shopping list with one item highlighted in bright green,
|
||
he or she will be more likely to remember the highlighted item than any of the others.
|
||
Additionally, in the following list of words – desk, chair, bed, table, chipmunk, dresser,
|
||
stool, couch – "chipmunk" will be remembered the most as it stands out against the other
|
||
words in its meaning.
|
||
|
||
There have been many studies that demonstrate and confirm the von Restorff effect in children
|
||
and young adults. Another study found that college-aged students performed better when trying
|
||
to remember an outstanding item in a list during an immediate memory-task whereas elderly
|
||
individuals did not remember it well, suggesting a difference in processing strategies
|
||
between the age groups.
|
||
|
||
In yet another study, although a significant von Restorff effect was produced amongst both
|
||
age groups when manipulating font color, it was found to be smaller in older adults than
|
||
younger adults. This too indicates that older people display lesser benefits for distinctive
|
||
information compared to younger people.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Von_Restorff_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias PICTURE_SUPERIORITY_EFFECT = new()
|
||
{
|
||
Id = new Guid("2b8f679b-480c-4588-96b5-951767f870e3"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Picture Superiority Effect",
|
||
Description =
|
||
"""
|
||
# Picture Superiority Effect
|
||
The picture superiority effect refers to the phenomenon in which pictures and images are more
|
||
likely to be remembered than are words. This effect has been demonstrated in numerous experiments
|
||
using different methods. It is based on the notion that "human memory is extremely sensitive to
|
||
the symbolic modality of presentation of event information". Explanations for the picture
|
||
superiority effect are not concrete and are still being debated, however an evolutionary
|
||
explanation is that sight has a long history stretching back millions of years and was
|
||
crucial to survival in the past, whereas reading is a relatively recent invention, and
|
||
specific cognitive processes, such as decoding symbols and linking them to meaning.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("eeca14c3-8710-4522-8991-81db170d7f8b"), // MODALITY_EFFECT
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Picture_superiority_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias SELF_REFERENCE_EFFECT = new()
|
||
{
|
||
Id = new Guid("302b0004-1f18-4ed0-8fc1-6396fc7e6dbe"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Self-Reference Effect",
|
||
Description =
|
||
"""
|
||
# Self-Reference Effect
|
||
The self-reference effect is a tendency for people to encode information differently depending on
|
||
whether they are implicated in the information. When people are asked to remember information when
|
||
it is related in some way to themselves, the recall rate can be improved.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Self-reference_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias ANCHORING_EFFECT = new()
|
||
{
|
||
Id = new Guid("fe94ab26-70bb-4682-b7ee-e2828e4b02bd"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Anchoring Effect",
|
||
Description =
|
||
"""
|
||
# Anchoring Effect
|
||
The anchoring effect is a psychological phenomenon in which an individual's judgments or decisions
|
||
are influenced by a reference point or "anchor" which can be completely irrelevant. Both numeric
|
||
and non-numeric anchoring have been reported in research. In numeric anchoring, once the value
|
||
of the anchor is set, subsequent arguments, estimates, etc. made by an individual may change
|
||
from what they would have otherwise been without the anchor. For example, an individual may
|
||
be more likely to purchase a car if it is placed alongside a more expensive model (the anchor).
|
||
Prices discussed in negotiations that are lower than the anchor may seem reasonable, perhaps
|
||
even cheap to the buyer, even if said prices are still relatively higher than the actual market
|
||
value of the car. Another example may be when estimating the orbit of Mars, one might start
|
||
with the Earth's orbit (365 days) and then adjust upward until they reach a value that seems
|
||
reasonable (usually less than 687 days, the correct answer).
|
||
|
||
The original description of the anchoring effect came from psychophysics. When judging stimuli
|
||
along a continuum, it was noticed that the first and last stimuli were used to compare the other
|
||
stimuli (this is also referred to as "end anchoring"). This was applied to attitudes by Sherif
|
||
et al. in their 1958 article "Assimilation and effects of anchoring stimuli on judgments".
|
||
|
||
Anchoring in negotiation
|
||
In the negotiation process anchoring serves to determine an accepted starting point for the
|
||
subsequent negotiations. As soon as one side states their first price offer, the (subjective)
|
||
anchor is set. The counterbid (counter-anchor) is the second-anchor.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Anchoring_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias CONSERVATISM_BIAS = new()
|
||
{
|
||
Id = new Guid("8533edf9-3117-48c5-8f78-efbd996911f0"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Conservatism Bias",
|
||
Description =
|
||
"""
|
||
# Conservatism Bias
|
||
In cognitive psychology and decision science, conservatism or conservatism bias is a bias which refers
|
||
to the tendency to revise one's belief insufficiently when presented with new evidence. This bias
|
||
describes human belief revision in which people over-weigh the prior distribution (base rate) and
|
||
under-weigh new sample evidence when compared to Bayesian belief-revision. In other words, people
|
||
update their prior beliefs as new evidence becomes available, but they do so more slowly than
|
||
they would if they used Bayes' theorem.
|
||
|
||
In finance, evidence has been found that investors under-react to corporate events, consistent
|
||
with conservatism. This includes announcements of earnings, changes in dividends, and stock splits.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("1de0de03-a2a7-4248-b004-4152d84a3c86"), // BASE_RATE_FALLACY
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Conservatism_(belief_revision)",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias CONTRAST_EFFECT = new()
|
||
{
|
||
Id = new Guid("bc69c14d-0f2d-47ce-b20f-836fae36beb6"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Contrast Effect",
|
||
Description =
|
||
"""
|
||
# Contrast Effect
|
||
A contrast effect is the enhancement or diminishment, relative to normal, of perception, cognition
|
||
or related performance as a result of successive (immediately previous) or simultaneous exposure
|
||
to a stimulus of lesser or greater value in the same dimension. (Here, normal perception, cognition
|
||
or performance is that which would be obtained in the absence of the comparison stimulus—i.e.,
|
||
one based on all previous experience.)
|
||
|
||
Perception example: A neutral gray target will appear lighter or darker than it does in isolation
|
||
when immediately preceded by, or simultaneously compared to, respectively, a dark gray or light
|
||
gray target.
|
||
|
||
Cognition example: A person will appear more or less attractive than that person does in isolation
|
||
when immediately preceded by, or simultaneously compared to, respectively, a less or more attractive
|
||
person.
|
||
|
||
Performance example: A laboratory rat will work faster, or slower, during a stimulus predicting a
|
||
given amount of reward when that stimulus and reward are immediately preceded by, or alternated with,
|
||
respectively, different stimuli associated with either a lesser or greater amount of reward.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Contrast_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias DISTINCTION_BIAS = new()
|
||
{
|
||
Id = new Guid("0c65fbf8-c74a-49a1-8a16-0e789bce9524"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Distinction Bias",
|
||
Description =
|
||
"""
|
||
# Distinction Bias
|
||
The tendency to view two options as more distinctive when evaluating them simultaneously than when
|
||
evaluating them separately.
|
||
|
||
For example, when televisions are displayed next to each other on the sales floor, the difference
|
||
in quality between two very similar, high-quality televisions may appear great. A consumer may pay
|
||
a much higher price for the higher-quality television, even though the difference in quality is
|
||
imperceptible when the televisions are viewed in isolation. Because the consumer will likely be
|
||
watching only one television at a time, the lower-cost television would have provided a similar
|
||
experience at a lower cost.
|
||
|
||
To avoid this bias, avoid comparing two jobs, or houses, directly. Instead, consider each job, or
|
||
house, individually and make an overall assessment of each one on its own, and then compare
|
||
assessments, which allows them to make a choice that accurately predicts future experience.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("593f2a10-46a6-471e-9ab3-86df740df6f2"), // LESS_IS_BETTER_EFFECT
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Distinction_bias",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias FOCUSING_EFFECT = new()
|
||
{
|
||
Id = new Guid("490f26a1-3b9b-4048-9488-8ba93b8bd8af"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Focusing Effect",
|
||
Description =
|
||
"""
|
||
# Focusing Effect
|
||
A cognitive bias that occurs when people place too much importance on only one aspect of an
|
||
evaluation, causing an error in accurately predicting the utility of a future outcome.
|
||
|
||
Example: It is sunnier in California therefore people must be more happy there. Or a job
|
||
that pays more money must be better.
|
||
""",
|
||
|
||
Related = [],
|
||
Links = [],
|
||
};
|
||
|
||
private static readonly Bias FRAMING_EFFECT = new()
|
||
{
|
||
Id = new Guid("a1950fc4-20e0-4d36-8e68-540b491b2d23"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Framing Effect",
|
||
Description =
|
||
"""
|
||
# Framing Effect
|
||
The framing effect is a cognitive bias in which people decide between options based on whether the options
|
||
are presented with positive or negative connotations. Individuals have a tendency to make risk-avoidant
|
||
or choices when options are positively framed, while selecting more loss-avoidant options when presented with
|
||
a negative frame. In studies of the bias, options are presented in terms of the probability of either
|
||
losses or gains. While differently expressed, the options described are in effect identical. Gain and
|
||
loss are defined in the scenario as descriptions of outcomes, for example, lives lost or saved, patients
|
||
treated or not treated, monetary gains or losses.
|
||
|
||
Prospect theory posits that a loss is more significant than the equivalent gain, that a sure gain (certainty
|
||
effect and pseudocertainty effect) is favored over a probabilistic gain, and that a probabilistic loss is
|
||
preferred to a definite loss. One of the dangers of framing effects is that people are often provided with
|
||
options within the context of only one of the two frames.
|
||
|
||
The concept helps to develop an understanding of frame analysis within social movements, and also in the
|
||
formation of political opinion where spin plays a large role in political opinion polls that are framed to
|
||
encourage a response beneficial to the organization that has commissioned the poll. It has been suggested
|
||
that the use of the technique is discrediting political polls themselves. The effect is reduced, or even
|
||
eliminated, if ample credible information is provided to people.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Framing_effect_(psychology)",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias MONEY_ILLUSION = new()
|
||
{
|
||
Id = new Guid("33136203-8d52-42e5-ad32-561b3c288676"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Money Illusion",
|
||
Description =
|
||
"""
|
||
# Money Illusion
|
||
In economics, money illusion, or price illusion, is a cognitive bias where money is thought of in nominal,
|
||
rather than real terms. In other words, the face value (nominal value) of money is mistaken for its
|
||
purchasing power (real value) at a previous point in time. Viewing purchasing power as measured by the
|
||
nominal value is false, as modern fiat currencies have no intrinsic value and their real value depends
|
||
purely on the price level. The term was coined by Irving Fisher in *Stabilizing the Dollar*. It was
|
||
popularized by John Maynard Keynes in the early twentieth century, and Irving Fisher wrote an important
|
||
book on the subject, *The Money Illusion*, in 1928.
|
||
|
||
The existence of money illusion is disputed by monetary economists who contend that people act rationally
|
||
(i.e. think in real prices) with regard to their wealth. Eldar Shafir, Peter A. Diamond, and Amos
|
||
Tversky (1997) have provided empirical evidence for the existence of the effect and it has been shown to
|
||
affect behaviour in a variety of experimental and real-world situations.
|
||
|
||
Shafir et al. also state that money illusion influences economic behaviour in three main ways:
|
||
|
||
- Price stickiness. Money illusion has been proposed as one reason why nominal prices are slow to change
|
||
even where inflation has caused real prices to fall or costs to rise.
|
||
|
||
- Contracts and laws are not indexed to inflation as frequently as one would rationally expect.
|
||
|
||
- Social discourse, in formal media and more generally, reflects some confusion about real and nominal value.
|
||
|
||
Money illusion can also influence people's perceptions of outcomes. Experiments have shown that people
|
||
generally perceive an approximate 2% cut in nominal income with no change in monetary value as unfair,
|
||
but see a 2% rise in nominal income where there is 4% inflation as fair, despite them being almost rational
|
||
equivalents. This result is consistent with the 'Myopic Loss Aversion theory'. Furthermore, the money illusion
|
||
means nominal changes in price can influence demand even if real prices have remained constant.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Money_illusion",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias WEBER_FECHNER_LAW = new()
|
||
{
|
||
Id = new Guid("528077d5-fdad-47df-89d4-6a32287c321b"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Weber–Fechner Law",
|
||
Description =
|
||
"""
|
||
# Weber–Fechner Law
|
||
The Weber–Fechner laws are two related scientific laws in the field of psychophysics, known as Weber's law and
|
||
Fechner's law. Both relate to human perception, more specifically the relation between the actual change in a
|
||
physical stimulus and the perceived change. This includes stimuli to all senses: vision, hearing, taste, touch,
|
||
and smell.
|
||
|
||
Ernst Heinrich Weber states that "the minimum increase of stimulus which will produce a perceptible increase
|
||
of sensation is proportional to the pre-existent stimulus," while Gustav Fechner's law is an inference from
|
||
Weber's law (with additional assumptions) which states that the intensity of our sensation increases as the
|
||
logarithm of an increase in energy rather than as rapidly as the increase.
|
||
|
||
Psychological studies show that it becomes increasingly difficult to discriminate between two numbers as the
|
||
difference between them decreases. This is called the distance effect. This is important in areas of magnitude
|
||
estimation, such as dealing with large scales and estimating distances. It may also play a role in explaining
|
||
why consumers neglect to shop around to save a small percentage on a large purchase, but will shop around to
|
||
save a large percentage on a small purchase which represents a much smaller absolute dollar amount.
|
||
|
||
Preliminary research has found that pleasant emotions adhere to Weber’s Law, with accuracy in judging their
|
||
intensity decreasing as pleasantness increases. However, this pattern wasn't observed for unpleasant emotions,
|
||
suggesting a survival-related need for accurately discerning high-intensity negative emotions.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Weber%E2%80%93Fechner_law",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias CONFIRMATION_BIAS = new()
|
||
{
|
||
Id = new Guid("af63ce77-f6c6-4e0f-8a9e-3daedc497f9a"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Confirmation Bias",
|
||
Description =
|
||
"""
|
||
# Confirmation Bias
|
||
Confirmation bias (also confirmatory bias, myside bias, or congeniality bias) is the tendency to search for,
|
||
interpret, favor, and recall information in a way that confirms or supports one's prior beliefs or values.
|
||
People display this bias when they select information that supports their views, ignoring contrary information,
|
||
or when they interpret ambiguous evidence as supporting their existing attitudes. The effect is strongest for
|
||
desired outcomes, for emotionally charged issues, and for deeply entrenched beliefs.
|
||
|
||
In social media, confirmation bias is amplified by the use of filter bubbles, or "algorithmic editing", which
|
||
displays to individuals only information they are likely to agree with, while excluding opposing views. Some
|
||
have argued that confirmation bias is the reason why society can never escape from filter bubbles, because
|
||
individuals are psychologically hardwired to seek information that agrees with their preexisting values and
|
||
beliefs. Others have further argued that the mixture of the two is degrading democracy—claiming that this
|
||
"algorithmic editing" removes diverse viewpoints and information—and that unless filter bubble algorithms
|
||
are removed, voters will be unable to make fully informed political decisions.
|
||
|
||
Many times in the history of science, scientists have resisted new discoveries by selectively interpreting or
|
||
ignoring unfavorable data. Several studies have shown that scientists rate studies that report findings
|
||
consistent with their prior beliefs more favorably than studies reporting findings inconsistent with
|
||
their previous beliefs. Further, confirmation biases can sustain scientific theories or research programs
|
||
in the face of inadequate or even contradictory evidence. The discipline of parapsychology is often cited
|
||
as an example.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("4e571eaf-7c2b-44c8-b8cb-0c8da658b82d"), // FREQUENCY_ILLUSION
|
||
new Guid("d749ce96-32f3-4c3d-86f7-26ff4edabe4a"), // AVAILABILITY_HEURISTIC
|
||
new Guid("0378a05c-b55b-4451-a7f4-b5e1d6287d83"), // FADING_AFFECT_BIAS
|
||
new Guid("fee14af4-34af-4cd0-a72c-9ad489516b60"), // CONGRUENCE_BIAS
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Confirmation_bias",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias CONGRUENCE_BIAS = new()
|
||
{
|
||
Id = new Guid("fee14af4-34af-4cd0-a72c-9ad489516b60"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Congruence Bias",
|
||
Description =
|
||
"""
|
||
# Congruence Bias
|
||
Congruence bias is the tendency of people to over-rely on testing their initial hypothesis (the most congruent one)
|
||
while neglecting to test alternative hypotheses. That is, people rarely try experiments that could disprove their
|
||
initial belief, but rather try to repeat their initial results. It is a special case of the confirmation bias.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("af63ce77-f6c6-4e0f-8a9e-3daedc497f9a"), // CONFIRMATION_BIAS
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Congruence_bias",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias CHOICE_SUPPORTIVE_BIAS = new()
|
||
{
|
||
Id = new Guid("51337702-9dc4-442a-8584-78f56e9ec186"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Choice-supportive Bias",
|
||
Description =
|
||
"""
|
||
# Choice-supportive Bias
|
||
Choice-supportive bias or post-purchase rationalization is the tendency to retroactively ascribe positive attributes
|
||
to an option one has selected and/or to demote the forgone options. It is part of cognitive science, and is a
|
||
distinct cognitive bias that occurs once a decision is made. For example, if a person chooses option A instead of
|
||
option B, they are likely to ignore or downplay the faults of option A while amplifying or ascribing new negative
|
||
faults to option B. Conversely, they are also likely to notice and amplify the advantages of option A and not notice
|
||
or de-emphasize those of option B.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("dd45c762-0599-4c6d-82e0-d10f7ee85bb1"), // MISATTRIBUTION_OF_MEMORY
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Choice-supportive_bias",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias SELECTIVE_PERCEPTION = new()
|
||
{
|
||
Id = new Guid("91fded4f-de89-405e-8627-dba49cf5deaa"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Selective Perception",
|
||
Description =
|
||
"""
|
||
# Selective Perception
|
||
Selective perception is the tendency not to notice and more quickly forget stimuli that cause emotional discomfort
|
||
and contradict prior beliefs. For example, a teacher may have a favorite student because they are biased by in-group
|
||
favoritism. The teacher ignores the student's poor attainment. Conversely, they might not notice the progress of
|
||
their least favorite student. It can also occur when consuming mass media, allowing people to see facts and
|
||
opinions they like while ignoring those that do not fit with particular opinions, values, beliefs, or frame of
|
||
reference. Psychologists believe this process occurs automatically.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("75e51ef5-f992-41c2-8778-0002c617db9a"), // OSTRICH_EFFECT
|
||
new Guid("1dfd3e9e-e44e-44cf-b8a0-95dea7a0e780"), // NORMALCY_BIAS
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Selective_perception",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias OBSERVER_EXPECTANCY_EFFECT = new()
|
||
{
|
||
Id = new Guid("c06c9a63-15aa-4601-aff4-ddfe6dd9727a"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Observer-Expectancy Effect",
|
||
Description =
|
||
"""
|
||
# Observer-Expectancy Effect
|
||
The observer-expectancy effect[a] is a form of reactivity in which a researcher's cognitive bias causes them to
|
||
subconsciously influence the participants of an experiment. Confirmation bias can lead to the experimenter
|
||
interpreting results incorrectly because of the tendency to look for information that conforms to their hypothesis,
|
||
and overlook information that argues against it. It is a significant threat to a study's internal validity, and
|
||
is therefore typically controlled using a double-blind experimental design.
|
||
|
||
The classic example of experimenter bias is that of "Clever Hans", an Orlov Trotter horse claimed by his owner
|
||
von Osten to be able to do arithmetic and other tasks. As a result of the large public interest in Clever Hans,
|
||
philosopher and psychologist Carl Stumpf, along with his assistant Oskar Pfungst, investigated these claims.
|
||
Ruling out simple fraud, Pfungst determined that the horse could answer correctly even when von Osten did not
|
||
ask the questions. However, the horse was unable to answer correctly when either it could not see the questioner,
|
||
or if the questioner themselves was unaware of the correct answer: When von Osten knew the answers to the questions,
|
||
Hans answered correctly 89% of the time. However, when von Osten did not know the answers, Hans guessed only 6% of
|
||
questions correctly. Pfungst then proceeded to examine the behaviour of the questioner in detail, and showed that
|
||
as the horse's taps approached the right answer, the questioner's posture and facial expression changed in ways
|
||
that were consistent with an increase in tension, which was released when the horse made the final, correct tap.
|
||
This provided a cue that the horse had learned to use as a reinforced cue to stop tapping.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("af63ce77-f6c6-4e0f-8a9e-3daedc497f9a"), // CONFIRMATION_BIAS
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Blinded_experiment",
|
||
"https://en.wikipedia.org/wiki/Observer-expectancy_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias OSTRICH_EFFECT = new()
|
||
{
|
||
Id = new Guid("75e51ef5-f992-41c2-8778-0002c617db9a"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Ostrich Effect",
|
||
Description =
|
||
"""
|
||
# Ostrich Effect
|
||
The ostrich effect, also known as the ostrich problem, was originally coined by Galai & Sade (2003). The name
|
||
comes from the common (but false) legend that ostriches bury their heads in the sand to avoid danger. This
|
||
effect is a cognitive bias where people tend to “bury their head in the sand” and avoid potentially negative
|
||
but useful information, such as feedback on progress, to avoid psychological discomfort.
|
||
|
||
There is neuroscientific evidence of the ostrich effect. Sharot et al. (2012) investigated the differences in
|
||
positive and negative information when updating existing beliefs. Consistent with the ostrich effect,
|
||
participants presented with negative information were more likely to avoid updating their beliefs.
|
||
|
||
An everyday example of the ostrich effect in a financial context is people avoiding checking their bank account
|
||
balance after spending a lot of money. There are known negative implications of the ostrich effect in healthcare.
|
||
For example, people with diabetes avoid monitoring their blood sugar levels.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("91fded4f-de89-405e-8627-dba49cf5deaa"), // SELECTIVE_PERCEPTION
|
||
new Guid("1dfd3e9e-e44e-44cf-b8a0-95dea7a0e780"), // NORMALCY_BIAS
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Ostrich_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias SUBJECTIVE_VALIDATION = new()
|
||
{
|
||
Id = new Guid("85612b34-0a78-454e-a204-7840bc11521c"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Subjective Validation",
|
||
Description =
|
||
"""
|
||
# Subjective Validation
|
||
Subjective validation, sometimes called personal validation effect, is a cognitive bias by which people will
|
||
consider a statement or another piece of information to be correct if it has any personal meaning or significance
|
||
to them. People whose opinion is affected by subjective validation will perceive two unrelated events (i.e., a
|
||
coincidence) to be related because their personal beliefs demand that they be related. Closely related to the
|
||
Forer effect, subjective validation is an important element in cold reading. It is considered to be the main
|
||
reason behind most reports of paranormal phenomena.
|
||
|
||
Example: Belief in a cold reading. Cold reading is a set of techniques used by mentalists, psychics, fortune-tellers,
|
||
and mediums. Without prior knowledge, a practiced cold-reader can quickly obtain a great deal of information by
|
||
analyzing the person's body language, age, clothing or fashion, hairstyle, gender, sexual orientation, religion,
|
||
ethnicity, level of education, manner of speech, place of origin, etc. during a line of questioning.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("05e4a15d-5c3e-42e9-88aa-bb40350d17e2"), // BARNUM_EFFECT
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Subjective_validation",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias BARNUM_EFFECT = new()
|
||
{
|
||
Id = new Guid("2cb8514a-c4a2-4cf6-aed7-72d7870ace84"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Barnum Effect",
|
||
Description =
|
||
"""
|
||
# Barnum Effect
|
||
The Barnum effect, also called the Forer effect or, less commonly, the Barnum–Forer effect, is a common psychological
|
||
phenomenon whereby individuals give high accuracy ratings to descriptions of their personality that supposedly are
|
||
tailored specifically to them, yet which are in fact vague and general enough to apply to a wide range of people.
|
||
This effect can provide a partial explanation for the widespread acceptance of some paranormal beliefs and practices,
|
||
such as astrology, fortune telling, aura reading, and some types of personality tests.
|
||
|
||
Example: Belief in a cold reading. Cold reading is a set of techniques used by mentalists, psychics, fortune-tellers,
|
||
and mediums. Without prior knowledge, a practiced cold-reader can quickly obtain a great deal of information by
|
||
analyzing the person's body language, age, clothing or fashion, hairstyle, gender, sexual orientation, religion,
|
||
ethnicity, level of education, manner of speech, place of origin, etc. during a line of questioning.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("85612b34-0a78-454e-a204-7840bc11521c"), // SUBJECTIVE_VALIDATION
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Barnum_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias CONTINUED_INFLUENCE_EFFECT = new()
|
||
{
|
||
Id = new Guid("7169c5e4-ca95-4568-b816-a36e2049993b"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Continued Influence Effect",
|
||
Description =
|
||
"""
|
||
# Continued Influence Effect
|
||
The tendency to believe previously learned misinformation even after it has been corrected. Misinformation can still
|
||
influence inferences one generates after a correction has occurred.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("2b69b071-6587-4ea1-a4f5-aee4e2fef43c"), // MISINFORMATION_EFFECT
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/List_of_cognitive_biases#Other_memory_biases",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias SEMMELWEIS_REFLEX = new()
|
||
{
|
||
Id = new Guid("48e2374a-9919-43eb-baa6-fc8c4f837d31"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Semmelweis Reflex",
|
||
Description =
|
||
"""
|
||
# Semmelweis Reflex
|
||
The Semmelweis reflex or "Semmelweis effect" is a metaphor for the reflex-like tendency to reject new evidence
|
||
or new knowledge because it contradicts established norms, beliefs, or paradigms.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("af63ce77-f6c6-4e0f-8a9e-3daedc497f9a"), // CONFIRMATION_BIAS
|
||
new Guid("7256f3f1-6650-4c45-bb85-36d81c9edd1a"), // AUTHORITY_BIAS
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Semmelweis_reflex",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias BIAS_BLIND_SPOT = new()
|
||
{
|
||
Id = new Guid("d8f01e8b-23c3-47da-979e-f18a3d4e104d"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Bias Blind Spot",
|
||
Description =
|
||
"""
|
||
# Bias Blind Spot
|
||
The bias blind spot is the cognitive bias of recognizing the impact of biases on the judgment of others, while failing
|
||
to see the impact of biases on one's own judgment. The term was created by Emily Pronin, a social psychologist from
|
||
Princeton University's Department of Psychology, with colleagues Daniel Lin and Lee Ross. The bias blind spot is named
|
||
after the visual blind spot. Most people appear to exhibit the bias blind spot. In a sample of more than 600 residents
|
||
of the United States, more than 85% believed they were less biased than the average American. Only one participant
|
||
believed that they were more biased than the average American. People do vary with regard to the extent to which
|
||
they exhibit the bias blind spot. This phenomenon has been successfully replicated and it appears that in general,
|
||
stronger personal free will beliefs are associated with bias blind spot. It appears to be a stable individual
|
||
difference that is measurable.
|
||
|
||
The bias blind spot appears to be a true blind spot in that it is unrelated to actual decision making ability.
|
||
Performance on indices of decision making competence are not related to individual differences in bias blind spot.
|
||
In other words, most people appear to believe that they are less biased than others, regardless of their actual
|
||
decision making ability.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("80f9b496-798a-4a1e-a426-815f23b8698e"), // INTROSPECTION_ILLUSION
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Bias_blind_spot",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias INTROSPECTION_ILLUSION = new()
|
||
{
|
||
Id = new Guid("80f9b496-798a-4a1e-a426-815f23b8698e"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Introspection Illusion",
|
||
Description =
|
||
"""
|
||
# Introspection Illusion
|
||
The introspection illusion is a cognitive bias in which people wrongly think they have direct insight into the origins
|
||
of their mental states, while treating others' introspections as unreliable. The illusion has been examined in
|
||
psychological experiments, and suggested as a basis for biases in how people compare themselves to others. These
|
||
experiments have been interpreted as suggesting that, rather than offering direct access to the processes underlying
|
||
mental states, introspection is a process of construction and inference, much as people indirectly infer others' mental
|
||
states from their behaviour.
|
||
|
||
When people mistake unreliable introspection for genuine self-knowledge, the result can be an illusion of superiority
|
||
over other people, for example when each person thinks they are less biased and less conformist than the rest of the
|
||
group. Even when experimental subjects are provided with reports of other subjects' introspections, in as detailed a
|
||
form as possible, they still rate those other introspections as unreliable while treating their own as reliable.
|
||
Although the hypothesis of an introspection illusion informs some psychological research, the existing evidence is
|
||
arguably inadequate to decide how reliable introspection is in normal circumstances.
|
||
|
||
The phrase "introspection illusion" was coined by Emily Pronin. Pronin describes the illusion as having four components:
|
||
|
||
- People give a strong weighting to introspective evidence when assessing themselves.
|
||
|
||
- They do not give such a strong weight when assessing others.
|
||
|
||
- People disregard their own behaviour when assessing themselves (but not others).
|
||
|
||
- Own introspections are more highly weighted than others. It is not just that people lack access to each other's
|
||
introspections: they regard only their own as reliable.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("f8fd4635-69b3-47be-8243-8c7c6749cae2"), // ILLUSORY_SUPERIORITY
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Introspection_illusion",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias ILLUSORY_SUPERIORITY = new()
|
||
{
|
||
Id = new Guid("f8fd4635-69b3-47be-8243-8c7c6749cae2"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Illusory Superiority",
|
||
Description =
|
||
"""
|
||
# Illusory Superiority
|
||
In social psychology, illusory superiority is a cognitive bias wherein people overestimate their own qualities and
|
||
abilities compared to others. Illusory superiority is one of many positive illusions, relating to the self, that
|
||
are evident in the study of intelligence, the effective performance of tasks and tests, and the possession of
|
||
desirable personal characteristics and personality traits. Overestimation of abilities compared to an objective
|
||
measure is known as the overconfidence effect.
|
||
|
||
The term "illusory superiority" was first used by the researchers Van Yperen and Buunk, in 1991. The phenomenon is
|
||
also known as the above-average effect, the superiority bias, the leniency error, the sense of relative superiority,
|
||
the primus inter pares effect, and the Lake Wobegon effect, named after the fictional town where all the children are
|
||
above average. The Dunning-Kruger effect is a form of illusory superiority shown by people on a task where their
|
||
level of skill is low.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("b821d449-64e5-4c0a-9d5a-3fda609a9b86"), // OVERCONFIDENCE_EFFECT
|
||
new Guid("b9c06da1-d2eb-4871-8159-a2a6d25e9eff"), // DUNNING_KRUGER_EFFECT
|
||
new Guid("80f9b496-798a-4a1e-a426-815f23b8698e"), // INTROSPECTION_ILLUSION
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Illusory_superiority",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias OVERCONFIDENCE_EFFECT = new()
|
||
{
|
||
Id = new Guid("b821d449-64e5-4c0a-9d5a-3fda609a9b86"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Overconfidence Effect",
|
||
Description =
|
||
"""
|
||
# Overconfidence Effect
|
||
The overconfidence effect is a well-established bias in which a person's subjective confidence in their
|
||
judgments is reliably greater than the objective accuracy of those judgments, especially when confidence
|
||
is relatively high.
|
||
|
||
The most common way in which overconfidence has been studied is by asking people how confident they are of
|
||
specific beliefs they hold or answers they provide. The data show that confidence systematically exceeds
|
||
accuracy, implying people are more sure that they are correct than they deserve to be.
|
||
|
||
The following is an incomplete list of events related or triggered by bias/overconfidence and a failing
|
||
(safety) culture:
|
||
|
||
- Chernobyl disaster
|
||
- Sinking of the Titanic
|
||
- Space Shuttle Challenger disaster
|
||
- Space Shuttle Columbia disaster
|
||
- Deepwater Horizon oil spill
|
||
- Titan submersible implosion
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("f8fd4635-69b3-47be-8243-8c7c6749cae2"), // ILLUSORY_SUPERIORITY
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Overconfidence_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias NAÏVE_CYNICISM = new()
|
||
{
|
||
Id = new Guid("5ae6f7ec-3be2-47ad-ad75-0ed114f97fe0"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Naïve Cynicism",
|
||
Description =
|
||
"""
|
||
# Naïve Cynicism
|
||
Naïve cynicism is a philosophy of mind, cognitive bias and form of psychological egoism that occurs when
|
||
people naïvely expect more egocentric bias in others than actually is the case.
|
||
|
||
The term was formally proposed by Justin Kruger and Thomas Gilovich and has been studied across a wide range
|
||
of contexts including: negotiations, group-membership, marriage, economics, government policy and more.
|
||
|
||
The theory of naïve cynicism can be described as:
|
||
|
||
- I am not biased.
|
||
- You are biased if you disagree with me.
|
||
- Your intentions/actions reflect your underlying egocentric biases.
|
||
|
||
As with naïve cynicism, the theory of naïve realism hinges on the acceptance of the following three beliefs:
|
||
|
||
- I am not biased.
|
||
- Reasonable people are not biased.
|
||
- All others are biased.
|
||
|
||
Naïve cynicism can be thought of as the counter to naïve realism, which is the belief that an individual
|
||
perceives the social world objectively while others perceive it subjectively.
|
||
|
||
It is important to discern that naïve cynicism is related to the notion that others have an egocentric bias
|
||
that motivates them to do things for their own self-interest rather than for altruistic reasons.
|
||
|
||
Both of these theories, however, relate to the extent that adults credit or discredit the beliefs or statements
|
||
of others.
|
||
|
||
Example: Cold War
|
||
The American reaction to a Russian SALT treaty during the Cold War is one well-known example of naïve cynicism
|
||
in history. Political leaders negotiating on behalf of the United States discredited the offer simply because
|
||
it was proposed by the Russian side.
|
||
|
||
Former U.S. congressman Floyd Spence indicates the use of naïve cynicism in this quote:
|
||
"I have had a philosophy for some time in regard to SALT, and it goes like this: the Russians will not accept
|
||
a SALT treaty that is not in their best interest, and it seems to me that if it is their best interests, it
|
||
can‘t be in our best interest."
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Na%C3%AFve_cynicism",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias NAÏVE_REALISM = new()
|
||
{
|
||
Id = new Guid("f0ad095e-8e9c-4bfb-855e-11fb5dd58cea"),
|
||
Category = BiasCategory.TOO_MUCH_INFORMATION,
|
||
Name = "Naïve Realism",
|
||
Description =
|
||
"""
|
||
# Naïve Realism
|
||
In social psychology, naïve realism is the human tendency to believe that we see the world around us objectively,
|
||
and that people who disagree with us must be uninformed, irrational, or biased.
|
||
|
||
Naïve realism provides a theoretical basis for several other cognitive biases, which are systematic errors when it
|
||
comes to thinking and making decisions. These include the false consensus effect, actor–observer bias, bias blind
|
||
spot, and fundamental attribution error, among others.
|
||
|
||
Lee Ross and fellow psychologist Andrew Ward have outlined three interrelated assumptions, or "tenets", that make up
|
||
naïve realism. They argue that these assumptions are supported by a long line of thinking in social psychology,
|
||
along with several empirical studies. According to their model, people:
|
||
|
||
- Believe that they see the world objectively and without bias.
|
||
|
||
- Expect that others will come to the same conclusions, so long as they are exposed to the same information and
|
||
interpret it in a rational manner.
|
||
|
||
- Assume that others who do not share the same views must be ignorant, irrational, or biased.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("bc0dc6d3-5115-4def-91ae-a38aebed185e"), // FALSE_CONSENSUS_EFFECT
|
||
new Guid("d8f01e8b-23c3-47da-979e-f18a3d4e104d"), // BIAS_BLIND_SPOT
|
||
new Guid("5da6dcf4-ed01-4e14-99b0-7a624b16cf17"), // ACTOR_OBSERVER_BIAS
|
||
new Guid("f1570784-f8ec-46fd-8bb8-763aef31a04a"), // FUNDAMENTAL_ATTRIBUTION_ERROR
|
||
new Guid("ca2d4f1f-924f-44ae-886b-19240cf2c8c0"), // ULTIMATE_ATTRIBUTION_ERROR
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Na%C3%AFve_realism_(psychology)",
|
||
],
|
||
};
|
||
|
||
#endregion
|
||
|
||
#region NOT_ENOUGH_MEANING
|
||
|
||
private static readonly Bias CONFABULATION = new()
|
||
{
|
||
Id = new Guid("2bbea096-a2a6-413f-85ce-32b5ae18669f"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Confabulation",
|
||
Description =
|
||
"""
|
||
# Confabulation
|
||
In psychology, confabulation is a memory error consisting of the production of fabricated, distorted, or misinterpreted
|
||
memories about oneself or the world. People who confabulate present with incorrect memories ranging from subtle inaccuracies
|
||
to surreal fabrications, and may include confusion or distortion in the temporal framing (timing, sequence or duration) of
|
||
memories. In general, they are very confident about their recollections, even when challenged with contradictory evidence.
|
||
Confabulation occurs when individuals mistakenly recall false information, without intending to deceive.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Confabulation",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias CLUSTERING_ILLUSION = new()
|
||
{
|
||
Id = new Guid("2c2ed1a8-aa4d-486d-a9b4-5a16ae9230c9"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Clustering Illusion",
|
||
Description =
|
||
"""
|
||
# Clustering Illusion
|
||
The clustering illusion is the tendency to erroneously consider the inevitable "streaks" or "clusters" arising in small
|
||
samples from random distributions to be non-random. The illusion is caused by a human tendency to underpredict the amount
|
||
of variability likely to appear in a small sample of random or pseudorandom data.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("7fce783e-2120-4aad-9805-2c2a2b937b7d"), // ILLUSION_OF_CONTROL
|
||
new Guid("465418ae-54b8-42ef-a29e-6ee9e9ffa769"), // INSENSITIVITY_TO_SAMPLE_SIZE
|
||
new Guid("61cd7e34-23be-43ef-ab97-8118cef7d23f"), // MONTE_CARLO_FALLACY
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Clustering_illusion",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias INSENSITIVITY_TO_SAMPLE_SIZE = new()
|
||
{
|
||
Id = new Guid("465418ae-54b8-42ef-a29e-6ee9e9ffa769"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Insensitivity to Sample Size",
|
||
Description =
|
||
"""
|
||
# Insensitivity to Sample Size
|
||
Insensitivity to sample size is a cognitive bias that occurs when people judge the probability of obtaining a sample statistic
|
||
without respect to the sample size. For example, in one study, subjects assigned the same probability to the likelihood of
|
||
obtaining a mean height of above six feet (183 cm) in samples of 10, 100, and 1,000 men. In other words, variation is more
|
||
likely in smaller samples, but people may not expect this.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("2c2ed1a8-aa4d-486d-a9b4-5a16ae9230c9"), // CLUSTERING_ILLUSION
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Insensitivity_to_sample_size",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias NEGLECT_OF_PROBABILITY = new()
|
||
{
|
||
Id = new Guid("44c6efd7-53f1-4d22-82fe-25e941390089"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Neglect of Probability",
|
||
Description =
|
||
"""
|
||
# Neglect of Probability
|
||
The neglect of probability, a type of cognitive bias, is the tendency to disregard probability when making a decision under
|
||
uncertainty and is one simple way in which people regularly violate the normative rules for decision making. Small risks are
|
||
typically either neglected entirely or hugely overrated. The continuum between the extremes is ignored. The term probability
|
||
neglect was coined by Cass Sunstein.
|
||
|
||
There are many related ways in which people violate the normative rules of decision making with regard to probability including
|
||
the hindsight bias, the neglect of prior base rates effect, and the gambler's fallacy. However, this bias is different, in that,
|
||
rather than incorrectly using probability, the actor disregards it.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("b9edd2f0-8503-4eb5-a4c3-369fcb318894"), // HINDSIGHT_BIAS
|
||
new Guid("1de0de03-a2a7-4248-b004-4152d84a3c86"), // BASE_RATE_FALLACY
|
||
new Guid("61cd7e34-23be-43ef-ab97-8118cef7d23f"), // MONTE_CARLO_FALLACY
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Neglect_of_probability",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias ANECDOTAL_FALLACY = new()
|
||
{
|
||
Id = new Guid("a448fe93-b176-4b5f-9498-f57f3f970a67"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Anecdotal Fallacy",
|
||
Description =
|
||
"""
|
||
# Anecdotal Fallacy
|
||
Misuse of anecdotal evidence is an informal fallacy and is sometimes referred to as the "person who" fallacy
|
||
("I know a person who..."; "I know of a case where..." etc.) which places undue weight on experiences of close
|
||
peers which may not be typical.
|
||
|
||
A common way anecdotal evidence becomes unscientific is through fallacious reasoning such as the "Post hoc ergo
|
||
propter hoc" fallacy, the human tendency to assume that if one event happens after another, then the first must
|
||
be the cause of the second. Another fallacy involves inductive reasoning. For instance, if an anecdote illustrates
|
||
a desired conclusion rather than a logical conclusion, it is considered a faulty or hasty generalization.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Anecdotal_evidence",
|
||
"https://en.wikipedia.org/wiki/Post_hoc_ergo_propter_hoc",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias ILLUSION_OF_VALIDITY = new()
|
||
{
|
||
Id = new Guid("8f68af8b-7b27-4697-bcf6-8bd4a5392a22"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Illusion of Validity",
|
||
Description =
|
||
"""
|
||
# Illusion of Validity
|
||
A cognitive bias in which a person overestimates his or her ability to interpret and predict accurately the outcome
|
||
when analyzing a set of data, in particular when the data analyzed show a very consistent pattern — that is, when the
|
||
data "tell" a coherent story. This effect persists even when the person is aware of all the factors that limit the
|
||
accuracy of his or her predictions, that is when the data and/or methods used to judge them lead to highly fallible
|
||
predictions.
|
||
|
||
Example: Subjects reported higher confidence in a prediction of the final grade point average of a student after
|
||
seeing a first-year record of consistent B’s than a first-year record of an even number of A’s and C’s. Consistent
|
||
patterns may be observed when input variables are highly redundant or correlated, which may increase subjective
|
||
confidence. However, a number of highly correlated inputs should not increase confidence much more than only one
|
||
of the inputs; instead higher confidence should be merited when a number of highly independent inputs show a
|
||
consistent pattern.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("ecfa5b28-3900-45ba-89c7-f8d995dfe406"), // WYSIATI
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Illusion_of_validity",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias WYSIATI = new()
|
||
{
|
||
Id = new Guid("ecfa5b28-3900-45ba-89c7-f8d995dfe406"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "WYSIATI (What You See Is All There Is)",
|
||
Description =
|
||
"""
|
||
# WYSIATI (What You See Is All There Is)
|
||
It solves a difficult problem by replacing it with a simpler problem that you know about. One problem does not solve
|
||
the other. The acronym WYSIATI stands for "What you see is all there is." It was coined by Nobel laureate Daniel
|
||
Kahneman in his book "Thinking, Fast and Slow." WYSIATI refers to the fact that we make decisions based on the
|
||
information we currently have. For example, when we meet an unknown person, we decide within seconds whether we
|
||
like the person or not.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Thinking,_Fast_and_Slow",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias MASKED_MAN_FALLACY = new()
|
||
{
|
||
Id = new Guid("5ddf8011-0ba2-4341-9e18-46178f8d4fbe"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Masked-Man Fallacy",
|
||
Description =
|
||
"""
|
||
# Masked-Man Fallacy
|
||
In philosophical logic, the masked-man fallacy (also known as the intensional fallacy or epistemic fallacy) is
|
||
committed when one makes an illicit use of Leibniz's law in an argument. Leibniz's law states that if A and B
|
||
are the same object, then A and B are indiscernible (that is, they have all the same properties). By modus tollens,
|
||
this means that if one object has a certain property, while another object does not have the same property, the two
|
||
objects cannot be identical. The fallacy is "epistemic" because it posits an immediate identity between a subject's
|
||
knowledge of an object with the object itself, failing to recognize that Leibniz's Law is not capable of accounting
|
||
for intensional contexts.
|
||
|
||
The name of the fallacy comes from the example:
|
||
|
||
- Premise 1: I know who Claus is.
|
||
- Premise 2: I do not know who the masked man is.
|
||
- Conclusion: Therefore, Claus is not the masked man.
|
||
|
||
The premises may be true and the conclusion false if Claus is the masked man and the speaker does not know that.
|
||
Thus the argument is a fallacious one.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Masked-man_fallacy",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias RECENCY_ILLUSION = new()
|
||
{
|
||
Id = new Guid("d0a79f6e-7786-4dd7-8a3f-62f167252171"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Recency Illusion",
|
||
Description =
|
||
"""
|
||
# Recency Illusion
|
||
The recency illusion is the belief or impression, on the part of someone who has only recently become aware of a
|
||
long-established phenomenon, that the phenomenon itself must be of recent origin. The term was coined by Arnold
|
||
Zwicky, a linguist at Stanford University who is primarily interested in examples involving words, meanings,
|
||
phrases, and grammatical constructions. However, use of the term is not restricted to linguistic phenomena:
|
||
Zwicky has defined it simply as, "the belief that things you have noticed only recently are in fact recent".
|
||
According to Zwicky, the illusion is caused by selective attention.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("91fded4f-de89-405e-8627-dba49cf5deaa"), // SELECTIVE_PERCEPTION
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Recency_illusion",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias GAMBLERS_FALLACY = new()
|
||
{
|
||
Id = new Guid("61cd7e34-23be-43ef-ab97-8118cef7d23f"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Gambler's Fallacy",
|
||
Description =
|
||
"""
|
||
# Gambler's Fallacy
|
||
he gambler's fallacy, also known as the Monte Carlo fallacy or the fallacy of the maturity of chances, is the belief
|
||
that, if an event (whose occurrences are independent and identically distributed) has occurred less frequently than
|
||
expected, it is more likely to happen again in the future (or vice versa). The fallacy is commonly associated with
|
||
gambling, where it may be believed, for example, that the next dice roll is more than usually likely to be six
|
||
because there have recently been fewer than the expected number of sixes.
|
||
|
||
In a study aimed at discovering if the negative autocorrelation that exists with the gambler's fallacy existed in
|
||
the decision made by U.S. asylum judges, results showed that after two successive asylum grants, a judge would be
|
||
5.5% less likely to approve a third grant.
|
||
|
||
In the decision making of loan officers, it can be argued that monetary incentives are a key factor in biased
|
||
decision making, rendering it harder to examine the gambler's fallacy effect. However, research shows that loan
|
||
officers who are not incentivised by monetary gain are 8% less likely to approve a loan if they approved one
|
||
for the previous client.
|
||
|
||
Several video games feature the use of loot boxes, a collection of in-game items awarded on opening with random
|
||
contents set by rarity metrics, as a monetization scheme. Since around 2018, loot boxes have come under scrutiny
|
||
from governments and advocates on the basis they are akin to gambling, particularly for games aimed at youth.
|
||
Some games use a special "pity-timer" mechanism, that if the player has opened several loot boxes in a row
|
||
without obtaining a high-rarity item, subsequent loot boxes will improve the odds of a higher-rate item drop.
|
||
This is considered to feed into the gambler's fallacy since it reinforces the idea that a player will eventually
|
||
obtain a high-rarity item (a win) after only receiving common items from a string of previous loot boxes.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("44c6efd7-53f1-4d22-82fe-25e941390089"), // NEGLECT_OF_PROBABILITY
|
||
new Guid("2c2ed1a8-aa4d-486d-a9b4-5a16ae9230c9"), // CLUSTERING_ILLUSION
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Gambler%27s_fallacy",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias HOT_HAND_FALLACY = new()
|
||
{
|
||
Id = new Guid("5fd14849-7041-42ee-976e-9a2b10522d29"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Hot Hand Fallacy",
|
||
Description =
|
||
"""
|
||
# Hot Hand Fallacy
|
||
The "hot hand" (also known as the "hot hand phenomenon" or "hot hand fallacy") is a phenomenon, previously
|
||
considered a cognitive social bias, that a person who experiences a successful outcome has a greater
|
||
chance of success in further attempts. The concept is often applied to sports and skill-based tasks
|
||
in general and originates from basketball, where a shooter is more likely to score if their previous
|
||
attempts were successful; i.e., while having the "hot hand.” While previous success at a task can indeed
|
||
change the psychological attitude and subsequent success rate of a player, researchers for many years
|
||
did not find evidence for a "hot hand" in practice, dismissing it as fallacious. However, later research
|
||
questioned whether the belief is indeed a fallacy. Some recent studies using modern statistical
|
||
analysis have observed evidence for the "hot hand" in some sporting activities; however, other recent
|
||
studies have not observed evidence of the "hot hand". Moreover, evidence suggests that only a small
|
||
subset of players may show a "hot hand" and, among those who do, the magnitude (i.e., effect size) of the
|
||
"hot hand" tends to be small.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Hot_hand",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias ILLUSORY_CORRELATION = new()
|
||
{
|
||
Id = new Guid("829d3178-8ebc-417c-b587-2ead31525327"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Illusory Correlation",
|
||
Description =
|
||
"""
|
||
# Illusory Correlation
|
||
In psychology, illusory correlation is the phenomenon of perceiving a relationship between variables
|
||
(typically people, events, or behaviors) even when no such relationship exists. A false association
|
||
may be formed because rare or novel occurrences are more salient and therefore tend to capture one's
|
||
attention. This phenomenon is one way stereotypes form and endure. Hamilton & Rose (1980) found that
|
||
stereotypes can lead people to expect certain groups and traits to fit together, and then to overestimate
|
||
the frequency with which these correlations actually occur. These stereotypes can be learned and perpetuated
|
||
without any actual contact occurring between the holder of the stereotype and the group it is about.
|
||
|
||
Example: A woman has her purse stolen by a person of a specific demographic. Henceforth, she keeps her
|
||
close purse each time she sees a similar person.
|
||
|
||
Example: A man holds the belief that people in urban environments tend to be rude. Therefore, when he
|
||
meets someone who is rude he assumes that the person lives in a city, rather than a rural area.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Illusory_correlation",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias PAREIDOLIA = new()
|
||
{
|
||
Id = new Guid("274cc868-df03-4fae-9dca-ccb07a66aeaf"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Pareidolia",
|
||
Description =
|
||
"""
|
||
# Pareidolia
|
||
Pareidolia is the tendency for perception to impose a meaningful interpretation on a nebulous stimulus,
|
||
usually visual, so that one detects an object, pattern, or meaning where there is none. Pareidolia is
|
||
a type of apophenia.
|
||
|
||
Common examples include perceived images of animals, faces, or objects in cloud formations; seeing faces
|
||
in inanimate objects; or lunar pareidolia like the Man in the Moon or the Moon rabbit. The concept of
|
||
pareidolia may extend to include hidden messages in recorded music played in reverse or at higher- or
|
||
lower-than-normal speeds, and hearing voices (mainly indistinct) or music in random noise, such as that
|
||
produced by air conditioners or by fans.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("6ab69dc8-6fcc-42c3-b190-90125a15b49f"), // APHOPHENIA
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Pareidolia",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias APOPHENIA = new()
|
||
{
|
||
Id = new Guid("6ab69dc8-6fcc-42c3-b190-90125a15b49f"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Apophenia",
|
||
Description =
|
||
"""
|
||
# Apophenia
|
||
Apophenia is the tendency to perceive meaningful connections between unrelated things. The term
|
||
(German: Apophänie from the Greek verb ἀποφαίνειν (apophaínein)) was coined by psychiatrist Klaus
|
||
Conrad in his 1958 publication on the beginning stages of schizophrenia. He defined it as "unmotivated
|
||
seeing of connections [accompanied by] a specific feeling of abnormal meaningfulness". He described the
|
||
early stages of delusional thought as self-referential over-interpretations of actual sensory
|
||
perceptions, as opposed to hallucinations. Apophenia has also come to describe a human propensity to
|
||
unreasonably seek definite patterns in random information, such as can occur in gambling.
|
||
|
||
Pareidolia is a type of apophenia involving the perception of images or sounds in random stimuli.
|
||
Gamblers may imagine that they see patterns in the numbers that appear in lotteries, card games, or
|
||
roulette wheels, where no such patterns exist. A common example of this is the gambler's fallacy.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("274cc868-df03-4fae-9dca-ccb07a66aeaf"), // PAREIDOLIA
|
||
new Guid("61cd7e34-23be-43ef-ab97-8118cef7d23f"), // MONTE_CARLO_FALLACY
|
||
new Guid("2c2ed1a8-aa4d-486d-a9b4-5a16ae9230c9"), // CLUSTERING_ILLUSION
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Apophenia",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias ANTHROPOMORPHISM = new()
|
||
{
|
||
Id = new Guid("70470097-52a8-4ea7-a85c-ed88ad1ed972"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Anthropomorphism",
|
||
Description =
|
||
"""
|
||
# Anthropomorphism
|
||
Anthropomorphism is the attribution of human traits, emotions, or intentions to non-human entities.
|
||
It is considered to be an innate tendency of human psychology. Personification is the related
|
||
attribution of human form and characteristics to abstract concepts such as nations, emotions, and
|
||
natural forces, such as seasons and weather. Both have ancient roots as storytelling and artistic
|
||
devices, and most cultures have traditional fables with anthropomorphized animals as characters.
|
||
People have also routinely attributed human emotions and behavioral traits to wild as well as
|
||
domesticated animals.
|
||
|
||
Anthropomorphism can be used to assist learning. Specifically, anthropomorphized words and
|
||
describing scientific concepts with intentionality can improve later recall of these concepts.
|
||
|
||
In people with depression, social anxiety, or other mental illnesses, emotional support animals
|
||
are a useful component of treatment partially because anthropomorphism of these animals can satisfy
|
||
the patients' need for social connection.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Anthropomorphism",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias GROUP_ATTRIBUTION_ERROR = new()
|
||
{
|
||
Id = new Guid("577e79e5-0a53-4c4c-a2ea-d039870bfbb9"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Group Attribution Error",
|
||
Description =
|
||
"""
|
||
# Group Attribution Error
|
||
The group attribution error refers to people's tendency to believe either
|
||
|
||
(a) the characteristics of an individual group member are reflective of
|
||
the group as a whole, or
|
||
|
||
(b) a group's decision outcome must reflect the preferences of individual
|
||
group members, even when external information is available suggesting otherwise.
|
||
|
||
The group attribution error shares an attribution bias analogous to the fundamental
|
||
attribution error. Rather than focusing on individual's behavior, it relies on group
|
||
outcomes and attitudes as its main basis for conclusions.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("f1570784-f8ec-46fd-8bb8-763aef31a04a"), // FUNDAMENTAL_ATTRIBUTION_ERROR
|
||
new Guid("efceb4b1-e19f-4997-9f96-1657bb269b2d"), // ATTRIBUTION_BIAS
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Group_attribution_error",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias ATTRIBUTION_BIAS = new()
|
||
{
|
||
Id = new Guid("efceb4b1-e19f-4997-9f96-1657bb269b2d"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Attribution Bias",
|
||
Description =
|
||
"""
|
||
# Attribution Bias
|
||
In psychology, an attribution bias or attributional errors is a cognitive bias that refers to the
|
||
systematic errors made when people evaluate or try to find reasons for their own and others' behaviors.
|
||
It refers to the systematic patterns of deviation from norm or rationality in judgment, often leading to
|
||
perceptual distortions, inaccurate assessments, or illogical interpretations of events and behaviors.
|
||
|
||
Attributions are the judgments and assumptions people make about why others behave a certain way. However,
|
||
these judgments may not always reflect the true situation. Instead of being completely objective, people
|
||
often make errors in perception that lead to skewed interpretations of social situations. Attribution
|
||
biases are present in everyday life. For example, when a driver cuts someone off, the person who has been
|
||
cut off is often more likely to attribute blame to the reckless driver's inherent personality traits (e.g.,
|
||
"That driver is rude and incompetent") rather than situational circumstances (e.g., "That driver may have
|
||
been late to work and was not paying attention").
|
||
|
||
Additionally, there are many different types of attribution biases, such as the ultimate attribution error,
|
||
fundamental attribution error, actor-observer bias, and hostile attribution bias. Each of these biases
|
||
describes a specific tendency that people exhibit when reasoning about the cause of different behaviors.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("ca2d4f1f-924f-44ae-886b-19240cf2c8c0"), // ULTIMATE_ATTRIBUTION_ERROR
|
||
new Guid("f1570784-f8ec-46fd-8bb8-763aef31a04a"), // FUNDAMENTAL_ATTRIBUTION_ERROR
|
||
new Guid("5da6dcf4-ed01-4e14-99b0-7a624b16cf17"), // ACTOR_OBSERVER_BIAS
|
||
new Guid("e85d8b16-5a36-4b63-af07-72c5188f089f"), // HOSTILE_ATTRIBUTION_BIAS
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Attribution_bias",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias HOSTILE_ATTRIBUTION_BIAS = new()
|
||
{
|
||
Id = new Guid("e85d8b16-5a36-4b63-af07-72c5188f089f"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Hostile Attribution Bias",
|
||
Description =
|
||
"""
|
||
# Hostile Attribution Bias
|
||
Hostile attribution bias, or hostile attribution of intent, is the tendency to interpret others' behaviors as
|
||
having hostile intent, even when the behavior is ambiguous or benign. For example, a person with high levels
|
||
of hostile attribution bias might see two people laughing and immediately interpret this behavior as two people
|
||
laughing about them, even though the behavior was ambiguous and may have been benign.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("efceb4b1-e19f-4997-9f96-1657bb269b2d"), // ATTRIBUTION_BIAS
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Hostile_attribution_bias",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias ULTIMATE_ATTRIBUTION_ERROR = new()
|
||
{
|
||
Id = new Guid("ca2d4f1f-924f-44ae-886b-19240cf2c8c0"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Ultimate Attribution Error",
|
||
Description =
|
||
"""
|
||
# Ultimate Attribution Error
|
||
The ultimate attribution error is a type of attribution error which describes how attributions of outgroup
|
||
behavior are more negative than ingroup behavior. As a cognitive bias, the error results in negative outgroup
|
||
behavior being more likely to be attributed to factors internal and specific to the actor, such as personality,
|
||
and the attribution of negative ingroup behavior to external factors such as luck or circumstance. The bias
|
||
reinforces negative stereotypes and prejudice about the outgroup and favouritism of the ingroup through positive
|
||
stereotypes. The theory also extends to the bias that positive acts performed by ingroup members are more likely
|
||
a result of their personality.
|
||
|
||
Four categories have been identified that describe the negative attribution of positive outgroup behaviour.
|
||
First, that the outgroup member is an exception to a general rule; second, that the member was lucky or had specific
|
||
advantages; third, that the member was highly motivated; and lastly that the behaviour as attributable to situational causes.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("efceb4b1-e19f-4997-9f96-1657bb269b2d"), // ATTRIBUTION_BIAS
|
||
new Guid("efb6606f-4629-4e5e-973f-94d5ac496638"), // PREJUDICE
|
||
new Guid("46c2a0b2-6b1b-4e02-86ea-3cff2bf292d0"), // STEREOTYPE
|
||
new Guid("b1cc861b-f445-450b-9bdf-e9d222abdb4e"), // IN_GROUP_FAVORITISM
|
||
new Guid("b57a862b-b490-4d61-96b8-29d548c2eee4"), // POSITIVITY_EFFECT
|
||
new Guid("f1570784-f8ec-46fd-8bb8-763aef31a04a"), // FUNDAMENTAL_ATTRIBUTION_ERROR
|
||
new Guid("5da6dcf4-ed01-4e14-99b0-7a624b16cf17"), // ACTOR_OBSERVER_BIAS
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Ultimate_attribution_error",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias IN_GROUP_FAVORITISM = new()
|
||
{
|
||
Id = new Guid("b1cc861b-f445-450b-9bdf-e9d222abdb4e"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "In-Group Favoritism",
|
||
Description =
|
||
"""
|
||
# In-Group Favoritism
|
||
In-group favoritism, sometimes known as in-group–out-group bias, in-group bias, intergroup bias, or
|
||
in-group preference, is a pattern of favoring members of one's in-group over out-group members.
|
||
This can be expressed in evaluation of others, in allocation of resources, and in many other ways.
|
||
|
||
This effect has been researched by many psychologists and linked to many theories related to group
|
||
conflict and prejudice. The phenomenon is primarily viewed from a social psychology standpoint.
|
||
Studies have shown that in-group favoritism arises as a result of the formation of cultural groups.
|
||
These cultural groups can be divided based on seemingly trivial observable traits, but with time,
|
||
populations grow to associate certain traits with certain behavior, increasing covariation. This
|
||
then incentivizes in-group bias.
|
||
|
||
Two prominent theoretical approaches to the phenomenon of in-group favoritism are realistic conflict
|
||
theory and social identity theory. Realistic conflict theory proposes that intergroup competition,
|
||
and sometimes intergroup conflict, arises when two groups have opposing claims to scarce resources.
|
||
In contrast, social identity theory posits a psychological drive for positively distinct social
|
||
identities as the general root cause of in-group favoring behavior.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("ca2d4f1f-924f-44ae-886b-19240cf2c8c0"), // ULTIMATE_ATTRIBUTION_ERROR
|
||
new Guid("6f5f4cbf-e6f3-439b-ad78-81b2dd266315"), // OUT_GROUP_HOMOGENEITY
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/In-group_favoritism",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias STEREOTYPING = new()
|
||
{
|
||
Id = new Guid("46c2a0b2-6b1b-4e02-86ea-3cff2bf292d0"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Stereotyping",
|
||
Description =
|
||
"""
|
||
# Stereotyping
|
||
In social psychology, a stereotype is a generalized belief about a particular category of people.
|
||
It is an expectation that people might have about every person of a particular group. The type of
|
||
expectation can vary; it can be, for example, an expectation about the group's personality, preferences,
|
||
appearance or ability. Stereotypes are often overgeneralized, inaccurate, and resistant to new information.
|
||
A stereotype does not necessarily need to be a negative assumption. They may be positive, neutral, or negative.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("ca2d4f1f-924f-44ae-886b-19240cf2c8c0"), // ULTIMATE_ATTRIBUTION_ERROR
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Stereotype",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias ESSENTIALISM = new()
|
||
{
|
||
Id = new Guid("179535d0-5da5-4c0f-b9b3-fb6644496254"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Essentialism",
|
||
Description =
|
||
"""
|
||
# Essentialism
|
||
The view that all objects have an essential substance that make the thing what it is, and without which
|
||
it would be not that kind of thing.
|
||
|
||
Essentialism has emerged as an important concept in psychology, particularly developmental psychology.
|
||
In 1991, Kathryn Kremer and Susan Gelman studied the extent to which children from four–seven years old
|
||
demonstrate essentialism. Children believed that underlying essences predicted observable behaviours.
|
||
Children were able to describe living objects' behaviour as self-perpetuated and non-living objects'
|
||
behavior as a result of an adult influencing the object. Understanding the underlying causal mechanism
|
||
for behaviour suggests essentialist thinking. Younger children were unable to identify causal
|
||
mechanisms of behaviour whereas older children were able to. This suggests that essentialism is rooted
|
||
in cognitive development. It can be argued that there is a shift in the way that children represent
|
||
entities, from not understanding the causal mechanism of the underlying essence to showing sufficient
|
||
understanding.
|
||
|
||
*Controversial. This is a philosophical viewpoint not a cognitive bias.*
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Essentialism",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias FUNCTIONAL_FIXEDNESS = new()
|
||
{
|
||
Id = new Guid("4346bdf9-4448-413f-92cd-4d146bf4789d"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Functional Fixedness",
|
||
Description =
|
||
"""
|
||
# Functional Fixedness
|
||
Functional fixedness is a cognitive bias that limits a person to use an object only in the way it is traditionally
|
||
used. The concept of functional fixedness originated in Gestalt psychology, a movement in psychology that emphasizes
|
||
holistic processing. Karl Duncker defined functional fixedness as being a mental block against using an object in a
|
||
new way that is required to solve a problem. This "block" limits the ability of an individual to use components given
|
||
to them to complete a task, as they cannot move past the original purpose of those components. For example, if someone
|
||
needs a paperweight, but they only have a hammer, they may not see how the hammer can be used as a paperweight.
|
||
Functional fixedness is this inability to see a hammer's use as anything other than for pounding nails; the person
|
||
couldn't think to use the hammer in a way other than in its conventional function.
|
||
|
||
When tested, 5-year-old children show no signs of functional fixedness. It has been argued that this is because at
|
||
age 5, any goal to be achieved with an object is equivalent to any other goal. However, by age 7, children have
|
||
acquired the tendency to treat the originally intended purpose of an object as special.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Functional_fixedness",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias MORAL_CREDENTIAL_EFFECT = new()
|
||
{
|
||
Id = new Guid("e36f82b7-43dd-4073-99d9-c33073007185"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Moral Credential Effect",
|
||
Description =
|
||
"""
|
||
# Moral Credential Effect
|
||
Self-licensing (also moral self-licensing, moral licensing, or licensing effect) is a term used in social
|
||
psychology and marketing to describe the subconscious phenomenon whereby increased confidence and security
|
||
in one's self-image or self-concept tends to make that individual worry less about the consequences of
|
||
subsequent immoral behavior and, therefore, more likely to make immoral choices and act immorally.
|
||
In simple terms, self-licensing occurs when people allow themselves to indulge after doing something positive
|
||
first; for example, drinking a diet soda with a greasy hamburger and fries can lead one to subconsciously
|
||
discount the negative attributes of the meal's high caloric and cholesterol content.
|
||
|
||
A large subset of this effect, the moral credential effect, is a bias that occurs when a person's track
|
||
record as a good egalitarian establishes in them an unconscious ethical certification, endorsement, or
|
||
license that increases the likelihood of less egalitarian decisions later. This effect occurs even when
|
||
the audience or moral peer group is unaware of the affected person's previously established moral credential.
|
||
For example, individuals who had the opportunity to recruit a woman or Black person in one setting were more
|
||
likely to say later, in a different setting, that a job would be better suited for a man or a white person.
|
||
Similar effects also appear to occur when a person observes another person from a group they identify with
|
||
making an egalitarian decision.
|
||
|
||
Self-licensing can have negative societal consequences since it has a permissive effect on behaviors such
|
||
as racial prejudice and discrimination, selfishness, poor dietary and health habits, and excessive
|
||
energy consumption.
|
||
|
||
But recent scholarship has failed to replicate seminal studies of the licensing effect, and meta-analysis
|
||
found it to be exaggerated by publication bias. Furthermore, where licensing typically assumes that a
|
||
good deed is the cause that makes subsequent transgressions more likely, an alternative (or additional)
|
||
account is that people are faced with a temptation to do something morally dubious, and use a prior good
|
||
deed as an excuse or reason why it is allowed for them to indulge.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Self-licensing",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias JUST_WORLD_FALLACY = new()
|
||
{
|
||
Id = new Guid("50c5f877-e656-494d-bc15-57c45a190cf9"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Just-World Fallacy",
|
||
Description =
|
||
"""
|
||
# Just-World Fallacy
|
||
The just-world fallacy, or just-world hypothesis, is the cognitive bias that assumes that "people get what
|
||
they deserve" – that actions will necessarily have morally fair and fitting consequences for the actor.
|
||
For example, the assumptions that noble actions will eventually be rewarded and evil actions will eventually
|
||
be punished fall under this fallacy. In other words, the just-world fallacy is the tendency to attribute
|
||
consequences to—or expect consequences as the result of— either a universal force that restores moral balance
|
||
or a universal connection between the nature of actions and their results. This belief generally implies the
|
||
existence of cosmic justice, destiny, divine providence, desert, stability, order, or the anglophone colloquial
|
||
use of "karma". It is often associated with a variety of fundamental fallacies, especially in regard to
|
||
rationalizing suffering on the grounds that the sufferers "deserve" it. This is called victim blaming.
|
||
|
||
This fallacy popularly appears in the English language in various figures of speech that imply guaranteed
|
||
punishment for wrongdoing, such as: "you got what was coming to you", "what goes around comes around",
|
||
"chickens come home to roost", "everything happens for a reason", and "you reap what you sow". This
|
||
hypothesis has been widely studied by social psychologists since Melvin J. Lerner conducted seminal work
|
||
on the belief in a just world in the early 1960s. Research has continued since then, examining the
|
||
predictive capacity of the fallacy in various situations and across cultures, and clarifying and expanding
|
||
the theoretical understandings of just-world beliefs.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Just-world_fallacy",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias ARGUMENT_FROM_FALLACY = new()
|
||
{
|
||
Id = new Guid("704695f1-9753-478b-9e9f-878e3a01e041"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Argument from Fallacy",
|
||
Description =
|
||
"""
|
||
# Argument from Fallacy
|
||
Argument from fallacy is the formal fallacy of analyzing an argument and inferring that, since it contains
|
||
a fallacy, its conclusion must be false. It is also called argument to logic (argumentum ad logicam), the
|
||
fallacy fallacy, the fallacist's fallacy, and the bad reasons fallacy.
|
||
|
||
Example 1:
|
||
- Alice: All cats are animals. Ginger is an animal. Therefore, Ginger is a cat.
|
||
- Bob: You have just fallaciously affirmed the consequent. You are incorrect. Therefore, Ginger is not a cat.
|
||
|
||
Example 2:
|
||
- Alice: I speak English. Therefore, I am English.
|
||
- Bob: Americans and Canadians, among others, speak English too. By assuming that speaking English and being
|
||
English always go together, you have just committed the package-deal fallacy. You are incorrect. Therefore,
|
||
you are not English.
|
||
|
||
Both of Bob's rebuttals are arguments from fallacy. Ginger may or may not be a cat, and Alice may or may not
|
||
be English. The fact that Alice's argument was fallacious is not, in itself, proof that her conclusion is false.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Argument_from_fallacy",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias AUTHORITY_BIAS = new()
|
||
{
|
||
Id = new Guid("7256f3f1-6650-4c45-bb85-36d81c9edd1a"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Authority Bias",
|
||
Description =
|
||
"""
|
||
# Authority Bias
|
||
Authority bias is the tendency to attribute greater accuracy to the opinion of an authority figure (unrelated to
|
||
its content) and be more influenced by that opinion. An individual is more influenced by the opinion of this
|
||
authority figure, believing their views to be more credible, and hence place greater emphasis on the authority
|
||
figure's viewpoint and are more likely to obey them. This concept is considered one of the social cognitive
|
||
biases or collective cognitive biases.
|
||
|
||
Cultural differences in the strength of authority bias have been identified, in which the differences in edits
|
||
made to Wikipedia articles by administrators and regular users were compared for accuracy. In Western Europe,
|
||
the bias has a negligible effect. In Eastern Europe, the bias is larger and the administrator's edits are
|
||
perceived as more likely to be true (despite the edits being inaccurate), indicating a cultural difference
|
||
in the extent to which authority bias is experienced.
|
||
|
||
Business: The authority bias is demonstrated in the case of the highest-paid persons' opinion (HIPPO) impact,
|
||
which describes how employees and other stakeholders in the solution environment tend to go with the opinions
|
||
and impressions of the highly paid people in an organization.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("af63ce77-f6c6-4e0f-8a9e-3daedc497f9a"), // CONFIRMATION_BIAS
|
||
new Guid("b1d46b0f-fa51-4e82-b0aa-71ba2c6ad1f1"), // BANDWAGON_EFFECT
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Authority_bias",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias AUTOMATION_BIAS = new()
|
||
{
|
||
Id = new Guid("c9e10d5b-6a32-4766-b937-aa03e276f018"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Automation Bias",
|
||
Description =
|
||
"""
|
||
# Automation Bias
|
||
Automation bias is the propensity for humans to favor suggestions from automated decision-making systems and to
|
||
ignore contradictory information made without automation, even if it is correct. Automation bias stems from
|
||
the social psychology literature that found a bias in human-human interaction that showed that people assign more
|
||
positive evaluations to decisions made by humans than to a neutral object. The same type of positivity bias
|
||
has been found for human-automation interaction, where the automated decisions are rated more positively
|
||
than neutral. This has become a growing problem for decision making as intensive care units, nuclear power
|
||
plants, and aircraft cockpits have increasingly integrated computerized system monitors and decision aids
|
||
to mostly factor out possible human error. Errors of automation bias tend to occur when decision-making
|
||
is dependent on computers or other automated aids and the human is in an observatory role but able to
|
||
make decisions. Examples of automation bias range from urgent matters like flying a plane on automatic
|
||
pilot to such mundane matters as the use of spell-checking programs.
|
||
|
||
An operator's trust in the system can also lead to different interactions with the system, including system
|
||
use, misuse, disuse, and abuse. Automation use and disuse can also influence stages of information processing:
|
||
information acquisition, information analysis, decision making and action selection, and action implementation.
|
||
|
||
For example, information acquisition, the first step in information processing, is the process by which a user
|
||
registers input via the senses. An automated engine gauge might assist the user with information acquisition
|
||
through simple interface features—such as highlighting changes in the engine's performance—thereby directing
|
||
the user's selective attention. When faced with issues originating from an aircraft, pilots may tend to
|
||
overtrust an aircraft's engine gauges, losing sight of other possible malfunctions not related to the engine.
|
||
This attitude is a form of automation complacency and misuse. If, however, the pilot devotes time to interpret
|
||
the engine gauge, and manipulate the aircraft accordingly, only to discover that the flight turbulence has not
|
||
changed, the pilot may be inclined to ignore future error recommendations conveyed by an engine gauge—a form
|
||
of automation complacency leading to disuse.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("b821d449-64e5-4c0a-9d5a-3fda609a9b86"), // OVERCONFIDENCE_EFFECT
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Automation_bias",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias BANDWAGON_EFFECT = new()
|
||
{
|
||
Id = new Guid("b1d46b0f-fa51-4e82-b0aa-71ba2c6ad1f1"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Bandwagon Effect",
|
||
Description =
|
||
"""
|
||
# Bandwagon Effect
|
||
The bandwagon effect is a psychological phenomenon where people adopt certain behaviors, styles, or attitudes
|
||
simply because others are doing so. More specifically, it is a cognitive bias by which public opinion or
|
||
behaviours can alter due to particular actions and beliefs rallying amongst the public. It is a psychological
|
||
phenomenon whereby the rate of uptake of beliefs, ideas, fads and trends increases with respect to the proportion
|
||
of others who have already done so. As more people come to believe in something, others also "hop on the bandwagon"
|
||
regardless of the underlying evidence.
|
||
|
||
Following others' actions or beliefs can occur because of conformism or deriving information from others. Much of
|
||
the influence of the bandwagon effect comes from the desire to 'fit in' with peers; by making similar selections
|
||
as other people, this is seen as a way to gain access to a particular social group. An example of this is fashion
|
||
trends wherein the increasing popularity of a certain garment or style encourages more acceptance. When individuals
|
||
make rational choices based on the information they receive from others, economists have proposed that information
|
||
cascades can quickly form in which people ignore their personal information signals and follow the behaviour of
|
||
others. Cascades explain why behaviour is fragile as people understand that their behaviour is based on a very
|
||
limited amount of information. As a result, fads form easily but are also easily dislodged. The phenomenon is
|
||
observed in various fields, such as economics, political science, medicine, and psychology. In social psychology,
|
||
people's tendency to align their beliefs and behaviors with a group is known as 'herd mentality' or 'groupthink'.
|
||
The reverse bandwagon effect (also known as the snob effect in certain contexts) is a cognitive bias that causes
|
||
people to avoid doing something, because they believe that other people are doing it.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("7256f3f1-6650-4c45-bb85-36d81c9edd1a"), // AUTHORITY_BIAS
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Bandwagon_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias PLACEBO_EFFECT = new()
|
||
{
|
||
Id = new Guid("8d76fae9-cd8e-46b5-9cbc-c8fffa6613a8"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Placebo Effect",
|
||
Description =
|
||
"""
|
||
# Placebo Effect
|
||
The psychological phenomenon in which the recipient perceives an improvement in condition due to personal
|
||
expectations rather than treatment itself.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Placebo",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias OUT_GROUP_HOMOGENEITY = new()
|
||
{
|
||
Id = new Guid("6f5f4cbf-e6f3-439b-ad78-81b2dd266315"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Out-Group Homogeneity",
|
||
Description =
|
||
"""
|
||
# Out-Group Homogeneity
|
||
The out-group homogeneity effect is the perception of out-group members as more similar to one another than are in-group members,
|
||
e.g. "they are alike; we are diverse". Perceivers tend to have impressions about the diversity or variability of group members around
|
||
those central tendencies or typical attributes of those group members. Thus, outgroup stereotypicality judgments are overestimated,
|
||
supporting the view that out-group stereotypes are overgeneralizations. The term "outgroup homogeneity effect", "outgroup homogeneity
|
||
bias" or "relative outgroup homogeneity" have been explicitly contrasted with "outgroup homogeneity" in general, the latter referring
|
||
to perceived outgroup variability unrelated to perceptions of the ingroup.
|
||
|
||
The outgroup homogeneity effect is sometimes referred to as "outgroup homogeneity bias". Such nomenclature hints at a broader
|
||
meta-theoretical debate that is present in the field of social psychology. This debate centres on the validity of heightened perceptions
|
||
of ingroup and outgroup homogeneity, where some researchers view the homogeneity effect as an example of cognitive bias and error, while
|
||
other researchers view the effect as an example of normal and often adaptive social perception. The out-group homogeneity effect has
|
||
been found using a wide variety of different social groups, from political and racial groups to age and gender groups.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("b1cc861b-f445-450b-9bdf-e9d222abdb4e"), // IN_GROUP_FAVORITISM
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Out-group_homogeneity",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias CROSS_RACE_EFFECT = new()
|
||
{
|
||
Id = new Guid("d36f046d-fe5c-4f4a-8d7f-14427b834581"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Cross-Race Effect",
|
||
Description =
|
||
"""
|
||
# Cross-Race Effect
|
||
The cross-race effect (sometimes called cross-race bias, other-race bias, own-race bias or other-race effect) is the tendency to more easily
|
||
recognize faces that belong to one's own racial group, or racial groups that one has been in contact with. In social psychology, the
|
||
cross-race effect is described as the "ingroup advantage," whereas in other fields, the effect can be seen as a specific form of the
|
||
"ingroup advantage" since it is only applied in interracial or inter-ethnic situations. The cross-race effect is thought to contribute
|
||
to difficulties in cross-race identification, as well as implicit racial bias.
|
||
|
||
A number of theories as to why the cross-race effect exists have been conceived, including social cognition and perceptual expertise.
|
||
However, no model has been able to fully account for the full body of evidence.
|
||
|
||
Cross-race identification bias
|
||
This effect refers to the decreased ability of people of one race to recognize faces and facial expressions of people of another race. This
|
||
differs from the cross-race bias because this effect is found mostly during eyewitness identification as well as identification of a suspect
|
||
in a line-up. In these situations, many people feel as if races other than their own look alike, and they have difficulty distinguishing
|
||
between members of different ethnic groups. Cross-race identification bias is also known as the misinformation effect since people are
|
||
considered to be misinformed about other races and have difficulty identifying them. A study was made which examined 271 real court
|
||
cases. In photographic line-ups, 231 witnesses participated in cross-race versus same-race identification. In cross-race lineups,
|
||
only 45% were correctly identified versus 60% for same-race identifications. In a study dealing with eyewitness testimony,
|
||
investigators examined forty participants in a racially diverse area of the US. Participants watched a video of a property crime
|
||
being committed, then in the next 24 hours came to pick the suspect out of a photo line-up. Most of the participants in the study
|
||
either misidentified the suspect or stated the suspect was not in the line-up at all. Correct identification of the suspect
|
||
occurred more often when the eyewitness and the suspect were of the same race. In another study, 86 convenience store
|
||
clerks were asked to identify three customers: one white, one black, and one Mexican, all of whom had purchased in the store
|
||
earlier that day. The clerks tended to identify customers belonging to their own race accurately, but were more likely to make
|
||
errors when attempting to identify other races members. Meanwhile, another study found that "alcohol intoxication reduces
|
||
the own-race bias in face recognition," albeit by impairing accurate perception and leaving in place or increasing random error
|
||
rather than by improving facial recognition of members of other groups.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Cross-race_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias CHEERLEADER_EFFECT = new()
|
||
{
|
||
Id = new Guid("79f705e9-c461-4ad7-8b5e-83358aa345f7"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Cheerleader Effect",
|
||
Description =
|
||
"""
|
||
# Cheerleader Effect
|
||
The cognitive bias which causes people to think individuals are more attractive when they are in a group. This effect occurs
|
||
with male-only, female-only and mixed gender groups; and both small and large groups. The effect occurs to the same extent
|
||
with groups of four and 16 people. Participants in studies looked more at the attractive people than the unattractive people
|
||
in the group. The effect does not occur because group photos give the impression that individuals have more social or emotional
|
||
intelligence. This was shown to be the case by a study which used individual photos grouped together in a single image, rather
|
||
than photos taken of people in a group. The study generated the same effect.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Cheerleader_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias POSITIVITY_EFFECT = new()
|
||
{
|
||
Id = new Guid("b57a862b-b490-4d61-96b8-29d548c2eee4"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Positivity Effect",
|
||
Description =
|
||
"""
|
||
# Positivity Effect
|
||
The positivity effect is the ability to constructively analyze a situation where the desired results are not achieved, but still
|
||
obtain positive feedback that assists one's future progression. Empirical research findings suggest that the positivity effect
|
||
can be influenced by internal positive speech, where engaging in constructive self-dialogue can significantly improve one’s
|
||
ability to perceive and react to challenging situations more optimistically.
|
||
|
||
The findings of a study show that the optimism bias in future-oriented thinking fulfils a self-improvement purpose while also
|
||
suggesting this bias probably reflects a common underpinning motivational process across various future-thinking domains,
|
||
either episodic or semantic.
|
||
|
||
## In attribution
|
||
The positivity effect as an attribution phenomenon relates to the habits and characteristics of people when evaluating
|
||
the causes of their behaviors. To positively attribute is to be open to attributing a person’s inherent disposition as
|
||
the cause of their positive behaviors, and the situations surrounding them as the potential cause of their negative
|
||
behaviors.
|
||
|
||
## In perception
|
||
Two studies by Emilio Ferrara have shown that, on online social networks like Twitter and Instagram, users prefer to share
|
||
positive news, and are emotionally affected by positive news more than twice as much as they are by negative news.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("ef521fbb-c20b-47c9-87f8-a571a06a03eb"), // NEGATIVITY_BIAS
|
||
new Guid("ca2d4f1f-924f-44ae-886b-19240cf2c8c0"), // ULTIMATE_ATTRIBUTION_ERROR
|
||
new Guid("f1570784-f8ec-46fd-8bb8-763aef31a04a"), // FUNDAMENTAL_ATTRIBUTION_ERROR
|
||
new Guid("5da6dcf4-ed01-4e14-99b0-7a624b16cf17"), // ACTOR_OBSERVER_BIAS
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Positivity_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias NOT_INVENTED_HERE = new()
|
||
{
|
||
Id = new Guid("72fd9f08-b3c2-40b7-8d56-a2e84d776041"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Not Invented Here",
|
||
Description =
|
||
"""
|
||
# Not Invented Here
|
||
Not invented here (NIH) is the tendency to avoid using or buying products, research, standards, or knowledge from external origins.
|
||
It is usually adopted by social, corporate, or institutional cultures. Research illustrates a strong bias against ideas from the
|
||
outside.
|
||
|
||
The reasons for not wanting to use the work of others are varied, but can include a desire to support a local economy instead of
|
||
paying royalties to a foreign license-holder, fear of patent infringement, lack of understanding of the foreign work, an
|
||
unwillingness to acknowledge or value the work of others, jealousy, belief perseverance, or forming part of a wider turf war.
|
||
As a social phenomenon, this tendency can manifest itself as an unwillingness to adopt an idea or product because it originates
|
||
from another culture, a form of tribalism and/or an inadequate effort in choosing the right approach for the business.
|
||
|
||
The term is typically used in a pejorative sense. The opposite predisposition is sometimes called "proudly found elsewhere" (PFE)
|
||
or "invented elsewhere".
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Not_invented_here",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias REACTIVE_DEVALUATION = new()
|
||
{
|
||
Id = new Guid("46493445-4a8b-4488-901c-85da417c80a3"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Reactive Devaluation",
|
||
Description =
|
||
"""
|
||
# Reactive Devaluation
|
||
Reactive devaluation is a cognitive bias that occurs when a proposal is devalued if it appears to originate from an antagonist. The
|
||
bias was proposed by Lee Ross and Constance Stillinger (1988). Reactive devaluation could be caused by loss aversion or attitude
|
||
polarization, or naïve realism.
|
||
|
||
In an initial experiment, Stillinger and co-authors asked pedestrians in the US whether they would support a drastic bilateral
|
||
nuclear arms reduction program. If they were told the proposal came from President Ronald Reagan, 90 percent said it would be
|
||
favorable or even-handed to the United States; if they were told the proposal came from a group of unspecified policy analysts,
|
||
80 percent thought it was favorable or even; but, if respondents were told it came from Mikhail Gorbachev only 44 percent thought
|
||
it was favorable or neutral to the United States.
|
||
|
||
In another experiment, a contemporaneous controversy at Stanford University led to the university divesting of South African
|
||
assets because of the apartheid regime. Students at Stanford were asked to evaluate the University's divestment plan before
|
||
it was announced publicly and after such. Proposals including the actual eventual proposal were valued more highly when they
|
||
were hypothetical.
|
||
|
||
In another study, experimenters showed Israeli participants a peace proposal which had been actually proposed by Israel. If
|
||
participants were told the proposal came from a Palestinian source, they rated it lower than if they were told (correctly)
|
||
the identical proposal came from the Israeli government. If participants identified as "hawkish" were told it came from a
|
||
"dovish" Israeli government, they believed it was relatively bad for their people and good for the other side, but not if
|
||
participants identified as "doves".
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("ad3ed908-c56e-411b-a130-8af8574ff67b"), // LOSS_AVERSION
|
||
new Guid("f0ad095e-8e9c-4bfb-855e-11fb5dd58cea"), // NAÏVE_REALISM
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Reactive_devaluation",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias WELL_TRAVELLED_ROAD_EFFECT = new()
|
||
{
|
||
Id = new Guid("9ee2b5b5-463c-4bca-af85-087683f89ab3"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Well-Travelled Road Effect",
|
||
Description =
|
||
"""
|
||
# Well-Travelled Road Effect
|
||
The well travelled road effect is a cognitive bias in which travellers will estimate the time taken to traverse routes differently
|
||
depending on their familiarity with the route. Frequently travelled routes are assessed as taking a shorter time than unfamiliar
|
||
routes. This effect creates errors when estimating the most efficient route to an unfamiliar destination, when one candidate
|
||
route includes a familiar route, whilst the other candidate route includes no familiar routes. The effect is most salient when
|
||
subjects are driving, but is still detectable for pedestrians and users of public transport. The effect has been observed for
|
||
centuries but was first studied scientifically in the 1980s and 1990s following from earlier "heuristics and biases" work
|
||
undertaken by Daniel Kahneman and Amos Tversky.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Well_travelled_road_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias MENTAL_ACCOUNTING = new()
|
||
{
|
||
Id = new Guid("9444923f-90c9-4269-a4dc-291513fa6d12"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Mental Accounting",
|
||
Description =
|
||
"""
|
||
# Mental Accounting
|
||
Mental accounting (or psychological accounting) is a model of consumer behaviour developed by Richard Thaler that attempts to describe the
|
||
process whereby people code, categorize and evaluate economic outcomes. Mental accounting incorporates the economic concepts of prospect
|
||
theory and transactional utility theory to evaluate how people create distinctions between their financial resources in the form of mental
|
||
accounts, which in turn impacts the buyer decision process and reaction to economic outcomes. People are presumed to make mental accounts
|
||
as a self control strategy to manage and keep track of their spending and resources. People budget money into mental accounts for savings
|
||
(e.g., saving for a home) or expense categories (e.g., gas money, clothing, utilities). People also are assumed to make mental accounts to
|
||
facilitate savings for larger purposes (e.g., a home or college tuition). Mental accounting can result in people demonstrating greater
|
||
loss aversion for certain mental accounts, resulting in cognitive bias that incentivizes systematic departures from consumer rationality.
|
||
Through an increased understanding of mental accounting differences in decision making based on different resources, and different
|
||
reactions based on similar outcomes can be greater understood.
|
||
|
||
As Thaler puts it, “All organizations, from General Motors down to single person households, have explicit and/or implicit accounting
|
||
systems. The accounting system often influences decisions in unexpected ways”.
|
||
|
||
A more proximal psychological mechanism through which mental accounting influences spending is through its influence on the pain of
|
||
paying that is associated with spending money from a mental account. Pain of paying is a negative affective response associated
|
||
with a financial loss. Prototypical examples are the unpleasant feeling that one experiences when watching the fare increase on a
|
||
taximeter or at the gas pump. When considering an expense, consumers appear to compare the cost of the expense to the size of an
|
||
account that it would deplete (e.g., numerator vs. denominator). A $30 t-shirt, for example, would be a subjectively larger
|
||
expense when drawn from $50 in one's wallet than $500 in one's checking account. The larger the fraction, the more pain of
|
||
paying the purchase appears to generate and the less likely consumers are to then exchange money for the good. Other evidence
|
||
of the relation between pain of paying and spending include the lower debt held by consumers who report experiencing a higher
|
||
pain of paying for the same goods and services than consumers who report experiencing less pain of paying.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("ad3ed908-c56e-411b-a130-8af8574ff67b"), // LOSS_AVERSION
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Mental_accounting",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias APPEAL_TO_POSSIBILITY = new()
|
||
{
|
||
Id = new Guid("73ca0caa-25e5-4edb-91d4-f375a773f82c"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Appeal to Probability",
|
||
Description =
|
||
"""
|
||
# Appeal to Probability
|
||
An appeal to probability (or appeal to possibility, also known as possibiliter ergo probabiliter, "possibly, therefore
|
||
probably") is the logical fallacy of taking something for granted because it is possibly the case. The fact that an
|
||
event is possible does not imply that the event is probable, nor that the event was realized.
|
||
|
||
A fallacious appeal to possibility:
|
||
|
||
- If it can happen (premise).
|
||
- It will happen. (invalid conclusion)
|
||
|
||
- Something can go wrong (premise).
|
||
- Therefore, something will go wrong (invalid conclusion).
|
||
|
||
- If I do not bring my umbrella (premise)
|
||
- It will rain. (invalid conclusion).
|
||
|
||
Murphy's law is a (typically deliberate, tongue-in-cheek) invocation of the fallacy.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Appeal_to_probability",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias NORMALCY_BIAS = new()
|
||
{
|
||
Id = new Guid("1dfd3e9e-e44e-44cf-b8a0-95dea7a0e780"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Normalcy Bias",
|
||
Description =
|
||
"""
|
||
# Normalcy Bias
|
||
Normalcy bias, or normality bias, is a cognitive bias which leads people to disbelieve or minimize threat warnings. Consequently,
|
||
individuals underestimate the likelihood of a disaster, when it might affect them, and its potential adverse effects. The normalcy
|
||
bias causes many people to prepare inadequately for natural disasters, market crashes, and calamities caused by human error. About
|
||
80% of people reportedly display normalcy bias during a disaster.
|
||
|
||
The normalcy bias can manifest in response to warnings about disasters and actual catastrophes. Such events can range in scale
|
||
from incidents such as traffic collisions to global catastrophic risk. The event may involve social constructionism phenomena
|
||
such as loss of money in market crashes, or direct threats to continuity of life: as in natural disasters like a tsunami or
|
||
violence in war.
|
||
|
||
Normalcy bias has also been called analysis paralysis, the ostrich effect, and by first responders, the negative panic. The
|
||
opposite of normalcy bias is overreaction, or worst-case scenario bias, in which small deviations from normality are dealt
|
||
with as signals of an impending catastrophe.
|
||
|
||
## Prevention
|
||
The negative effects of normalcy bias can be combated through the four stages of disaster response:
|
||
|
||
- preparation, including publicly acknowledging the possibility of disaster and forming contingency plans.
|
||
|
||
- warning, including issuing clear, unambiguous, and frequent warnings and helping the public to understand and believe them.
|
||
|
||
- impact, the stage at which the contingency plans take effect and emergency services, rescue teams, and disaster relief
|
||
teams work in tandem.
|
||
|
||
- aftermath, reestablishing equilibrium after the fact, by providing both supplies and aid to those in need.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("75e51ef5-f992-41c2-8778-0002c617db9a"), // OSTRICH_EFFECT
|
||
new Guid("91fded4f-de89-405e-8627-dba49cf5deaa"), // SELECTIVE_PERCEPTION
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Normalcy_bias",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias ZERO_SUM_BIAS = new()
|
||
{
|
||
Id = new Guid("35c21723-8dd7-4fea-9404-b26660fa6db1"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Zero-Sum Bias",
|
||
Description =
|
||
"""
|
||
# Zero-Sum Thinking & Zero-Sum Bias
|
||
Zero-sum thinking perceives situations as zero-sum games, where one person's gain would be another's loss. The term is
|
||
derived from game theory. However, unlike the game theory concept, zero-sum thinking refers to a psychological
|
||
construct — a person's subjective interpretation of a situation. Zero-sum thinking is captured by the saying
|
||
"your gain is my loss" (or conversely, "your loss is my gain").
|
||
|
||
Rozycka-Tran et al. (2015) defined zero-sum thinking as:
|
||
"A general belief system about the antagonistic nature of social relations, shared by people in a society or culture
|
||
and based on the implicit assumption that a finite amount of goods exists in the world, in which one person's winning
|
||
makes others the losers, and vice versa ... a relatively permanent and general conviction that social relations are
|
||
like a zero-sum game. People who share this conviction believe that success, especially economic success, is possible
|
||
only at the expense of other people's failures."
|
||
|
||
Zero-sum bias is a cognitive bias towards zero-sum thinking; it is people's tendency to intuitively judge that a
|
||
situation is zero-sum, even when this is not the case. This bias promotes zero-sum fallacies, false beliefs that
|
||
situations are zero-sum. Such fallacies can cause other false judgements and poor decisions. In economics,
|
||
"zero-sum fallacy" generally refers to the fixed-pie fallacy.
|
||
|
||
## Examples
|
||
There are many examples of zero-sum thinking, some of them fallacious.
|
||
|
||
- When jurors assume that any evidence compatible with more than one theory offers no support for any theory, even
|
||
if the evidence is incompatible with some possibilities or the theories are not mutually exclusive.
|
||
|
||
- When students in a classroom think they are being graded on a curve when in fact they are being graded based
|
||
on predetermined standards.
|
||
|
||
- In a negotiation when one negotiator thinks that they can only gain at the expense of the other party (i.e.,
|
||
that mutual gain is not possible).
|
||
|
||
- In the context of social group competition, the belief that more resources for one group (e.g., immigrants)
|
||
means less for others (e.g., non-immigrants).
|
||
|
||
- Jack of all trades, master of none: the idea that having more skills means having less aptitude (also known
|
||
as compensatory reasoning).
|
||
|
||
- In copyright infringement debate, the idea that every unauthorized duplication is a lost sale.
|
||
|
||
- When politicians argue that international trade must mean that one party is "winning" and another is "losing"
|
||
when transfer of goods and services at mutually-agreeable prices is in general mutually beneficial, or that a
|
||
trade deficit represents "losing" money to another country.
|
||
|
||
- Group membership is sometimes treated as zero-sum, such that stronger membership in one group is seen as
|
||
weaker membership in another.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Zero-sum_thinking",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias SURVIVORSHIP_BIAS = new()
|
||
{
|
||
Id = new Guid("87ef31b2-6b2a-4fbb-9974-fefec5480c28"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Survivorship Bias",
|
||
Description =
|
||
"""
|
||
# Survivorship Bias
|
||
Survivorship bias or survival bias is the logical error of concentrating on entities that passed a selection
|
||
process while overlooking those that did not. This can lead to incorrect conclusions because of incomplete data.
|
||
|
||
Survivorship bias is a form of selection bias that can lead to overly optimistic beliefs because multiple
|
||
failures are overlooked, such as when companies that no longer exist are excluded from analyses of financial
|
||
performance. It can also lead to the false belief that the successes in a group have some special property,
|
||
rather than just coincidence as in correlation "proves" causality.
|
||
|
||
Another kind of survivorship bias would involve thinking that an incident happened in a particular way when
|
||
the only people who were involved in the incident who can speak about it are those who survived it. Even if
|
||
one knew that some people are dead, they would not have their voice to add to the conversation, making it
|
||
biased.
|
||
|
||
## Examples
|
||
### Finance and Economics
|
||
In finance, survivorship bias is the tendency for failed companies to be excluded from performance studies
|
||
because they no longer exist. It often causes the results of studies to skew higher because only companies
|
||
that were successful enough to survive until the end of the period are included. For example, a mutual fund
|
||
company's selection of funds today will include only those that are successful now. Many losing funds are
|
||
closed and merged into other funds to hide poor performance. In theory, 70% of extant funds could truthfully
|
||
claim to have performance in the first quartile of their peers, if the peer group includes funds that have
|
||
closed.
|
||
|
||
### Business
|
||
Michael Shermer in Scientific American and Larry Smith of the University of Waterloo have described
|
||
how advice about commercial success distorts perceptions of it by ignoring all of the businesses and college
|
||
dropouts that failed. Journalist and author David McRaney observes that the "advice business is a monopoly
|
||
run by survivors. When something becomes a non-survivor, it is either completely eliminated, or whatever
|
||
voice it has is muted to zero". Alec Liu wrote in Vice that "for every Mark Zuckerberg, there's
|
||
thousands of also-rans, who had parties no one ever attended, obsolete before we ever knew they
|
||
existed."
|
||
|
||
In his book The Black Swan, financial writer Nassim Taleb called the data obscured by survivorship bias
|
||
"silent evidence".
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Survivorship_bias",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias SUBADDITIVITY_EFFECT = new()
|
||
{
|
||
Id = new Guid("73e39503-4a2e-4090-88c2-5ce20565a722"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Subadditivity Effect",
|
||
Description =
|
||
"""
|
||
# Subadditivity Effect
|
||
The subadditivity effect is the tendency to judge probability of the whole to be less than the
|
||
probabilities of the parts.
|
||
|
||
Example:
|
||
For instance, subjects in one experiment judged the probability of death from cancer in the United
|
||
States was 18%, the probability from heart attack was 22%, and the probability of death from
|
||
"other natural causes" was 33%. Other participants judged the probability of death from a natural
|
||
cause was 58%. Natural causes are made up of precisely cancer, heart attack, and "other natural
|
||
causes," however, the sum of the latter three probabilities was 73%, and not 58%. According to
|
||
Tversky and Koehler (1994) this kind of result is observed consistently.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Subadditivity_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias DENOMINATION_EFFECT = new()
|
||
{
|
||
Id = new Guid("a913b2cf-dc2f-4dd9-87dc-3e11efb9457b"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Denomination Effect",
|
||
Description =
|
||
"""
|
||
# Denomination Effect
|
||
The denomination effect is a form of cognitive bias relating to currency, suggesting people may be
|
||
less likely to spend larger currency denominations than their equivalent value in smaller denominations.
|
||
It was proposed by Priya Raghubir, professor at the New York University Stern School of Business, and
|
||
Joydeep Srivastava, professor at University of Maryland, in their 2009 paper "Denomination Effect".
|
||
|
||
Raghubir and Srivastava conducted three studies in their research on the denomination effect; their
|
||
findings suggested people may be more likely to spend money represented by smaller denominations and
|
||
that consumers may prefer to receive money in a large denomination when there is a need to control
|
||
spending. The denomination effect can occur when large denominations are perceived as less exchangeable
|
||
than smaller denominations.
|
||
|
||
The effect's influence on spending decisions has implications throughout various sectors in society,
|
||
including consumer welfare, monetary policy and the finance industry. For example, during the Great
|
||
Recession, one businessman observed employees using more coins rather than banknotes in an office
|
||
vending machine, perceiving the customers used coins to feel thriftier. Raghubir and Srivastava
|
||
also suggested the effect may involve incentives to alter future behavior and that a large
|
||
denomination can serve as a mechanism to prevent the urge to spend.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("9444923f-90c9-4269-a4dc-291513fa6d12"), // MENTAL_ACCOUNTING
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Denomination_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias MILLERS_LAW = new()
|
||
{
|
||
Id = new Guid("81ca1f50-aaf9-4416-a94a-3676b26e510a"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Miller's Law",
|
||
Description =
|
||
"""
|
||
# Miller's Law
|
||
The observation, also by George A. Miller, that the number of objects the average person can hold in
|
||
working memory is about seven. It was put forward in a 1956 edition of Psychological Review in a
|
||
paper titled "The Magical Number Seven, Plus or Minus Two".
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Miller%27s_law",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias ILLUSION_OF_TRANSPARENCY = new()
|
||
{
|
||
Id = new Guid("c727e47c-da6f-4804-a1d0-9027af645218"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Illusion of Transparency",
|
||
Description =
|
||
"""
|
||
# Illusion of Transparency
|
||
The illusion of transparency is a tendency for people to overestimate the degree to which their personal
|
||
mental state is known by others. Another manifestation of the illusion of transparency (sometimes called
|
||
the observer's illusion of transparency) is a tendency for people to overestimate how well they understand
|
||
others' personal mental states. This cognitive bias is similar to the illusion of asymmetric insight.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("a44a6bcf-b2b8-47f1-84e0-d740af56aa1e"), // ILLUSION_OF_ASYMMETRIC_INSIGHT
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Illusion_of_transparency",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias CURSE_OF_KNOWLEDGE = new()
|
||
{
|
||
Id = new Guid("697f58a7-45d7-4268-8951-81681fb005de"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Curse of Knowledge",
|
||
Description =
|
||
"""
|
||
# Curse of Knowledge
|
||
The curse of knowledge, also called the curse of expertise or expert's curse, is a cognitive bias
|
||
that occurs when a person who has specialized knowledge assumes that others share in that knowledge.
|
||
|
||
For example, in a classroom setting, teachers may have difficulty if they cannot put themselves
|
||
in the position of the student. A knowledgeable professor might no longer remember the difficulties
|
||
that a young student encounters when learning a new subject for the first time. This curse of
|
||
knowledge also explains the danger behind thinking about student learning based on what appears
|
||
best to faculty members, as opposed to what has been verified with students.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Curse_of_knowledge",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias SPOTLIGHT_EFFECT = new()
|
||
{
|
||
Id = new Guid("1a6f6356-6d61-4892-8494-0257a7fa718b"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Spotlight Effect",
|
||
Description =
|
||
"""
|
||
# Spotlight Effect
|
||
The spotlight effect is the psychological phenomenon by which people tend to believe they are being
|
||
noticed more than they really are. Being that one is constantly in the center of one's own world,
|
||
an accurate evaluation of how much one is noticed by others is uncommon. The reason for the spotlight
|
||
effect is the innate tendency to forget that although one is the center of one's own world, one is
|
||
not the center of everyone else's. This tendency is especially prominent when one does something
|
||
atypical.
|
||
|
||
Research has empirically shown that such drastic over-estimation of one's effect on others is widely
|
||
common. Many professionals in social psychology encourage people to be conscious of the spotlight
|
||
effect and to allow this phenomenon to moderate the extent to which one believes one is in a
|
||
social spotlight.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Spotlight_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias EXTRINSIC_INCENTIVE_BIAS = new()
|
||
{
|
||
Id = new Guid("07237744-843d-4c0c-81b5-0c9c8664daea"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Extrinsic Incentives Bias",
|
||
Description =
|
||
"""
|
||
# Extrinsic Incentives Bias
|
||
The extrinsic incentives bias is an attributional bias according to which people attribute relatively
|
||
more to "extrinsic incentives" (such as monetary reward) than to "intrinsic incentives" (such as
|
||
learning a new skill) when weighing the motives of others rather than themselves.
|
||
|
||
It is a counter-example to the fundamental attribution error as according to the extrinsic bias
|
||
others are presumed to have situational motivations while oneself is seen as having dispositional
|
||
motivations. This is the opposite of what the fundamental attribution error would predict. It also
|
||
might help to explain some of the backfiring effects that can occur when extrinsic incentives are
|
||
attached to activities that people are intrinsically motivated to do. The term was first proposed
|
||
by Chip Heath, citing earlier research by others in management science.
|
||
|
||
Example:
|
||
In the simplest experiment Heath reported, MBA students were asked to rank the expected job motivations
|
||
of Citibank customer service representatives. Their average ratings were as follows:
|
||
|
||
1. Amount of pay
|
||
2. Having job security
|
||
3. Quality of fringe benefits
|
||
4. Amount of praise from your supervisor
|
||
5. Doing something that makes you feel good about yourself
|
||
6. Developing skills and abilities
|
||
7. Accomplishing something worthwhile
|
||
8. Learning new things
|
||
|
||
Actual customer service representatives rank ordered their own motivations as follows:
|
||
|
||
1. Developing skills and abilities
|
||
2. Accomplishing something worthwhile
|
||
3. Learning new things
|
||
4. Quality of fringe benefits
|
||
5. Having job security
|
||
6. Doing something that makes you feel good about yourself
|
||
7. Amount of pay
|
||
8. Amount of praise from your supervisor
|
||
|
||
The order of the predicted and actual reported motivations was nearly reversed; in particular, pay was
|
||
rated first by others but near last for respondents of themselves. Similar effects were observed when
|
||
MBA students rated managers and their classmates.
|
||
|
||
Debiasing:
|
||
Heath suggests trying to infer others' motivations as one would by inferring one's own motivations.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("f1570784-f8ec-46fd-8bb8-763aef31a04a"), // FUNDAMENTAL_ATTRIBUTION_ERROR
|
||
new Guid("bf8f304d-2e8e-4a90-a9c5-7bd56f6058a6"), // BACKFIRE_EFFECT
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Extrinsic_incentives_bias",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias ILLUSION_OF_EXTERNAL_AGENCY = new()
|
||
{
|
||
Id = new Guid("184c9dc0-6885-4dee-b777-bc1725cc7e2c"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Illusion of External Agency",
|
||
Description =
|
||
"""
|
||
# Illusion of External Agency
|
||
People typically underestimate their capacity to generate satisfaction with future outcomes. When people
|
||
experience such self-generated satisfaction, they may mistakenly conclude that it was caused by an
|
||
influential, insightful, and benevolent external agent.
|
||
|
||
When outcomes are unchangeable, people are more likely to turn ‘truly mediocre’ into ‘falsely great’.
|
||
This subjective transformation is often termed a psychological immune response, in that it is our brain
|
||
kicking in to protect us from the emotional consequences of undesirable outcomes. The illusion of external
|
||
agency is thought to arise from this undetected transformation of ‘truly mediocre’ outcomes to ‘falsely
|
||
great’ ones.
|
||
""",
|
||
|
||
Related = [],
|
||
Links = [],
|
||
};
|
||
|
||
private static readonly Bias ILLUSION_OF_ASYMMETRIC_INSIGHT = new()
|
||
{
|
||
Id = new Guid("a44a6bcf-b2b8-47f1-84e0-d740af56aa1e"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Illusion of Asymmetric Insight",
|
||
Description =
|
||
"""
|
||
# Illusion of Asymmetric Insight
|
||
The illusion of asymmetric insight is a cognitive bias whereby people perceive their knowledge of others to
|
||
surpass other people's knowledge of them. This bias "has been traced to people's tendency to view their own
|
||
spontaneous or off-the-cuff responses to others' questions as relatively unrevealing even though they view
|
||
others' similar responses as meaningful".
|
||
|
||
A study finds that people seem to believe that they know themselves better than their peers know themselves
|
||
and that their social group knows and understands other social groups better than other social groups know
|
||
them. For example: Person A knows Person A better than Person B knows Person B or Person A. This bias may be
|
||
sustained by a few cognitive beliefs, including:
|
||
|
||
- The personal conviction that observed behaviors are more revealing of other people than of the self, while
|
||
private thoughts and feelings are more revealing of the self.
|
||
|
||
- The more an individual perceives negative traits ascribed to someone else, the more doubt individuals express
|
||
about this person's self-knowledge. But, this doubt does not exist for our own self-knowledge. (For example:
|
||
if Person A believes Person B has some great character flaw, Person A will distrust Person B's self-knowledge,
|
||
while sustaining that they do not hold that same flaw in self-knowledge.)
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("c727e47c-da6f-4804-a1d0-9027af645218"), // ILLUSION_OF_TRANSPARENCY
|
||
new Guid("f1570784-f8ec-46fd-8bb8-763aef31a04a"), // FUNDAMENTAL_ATTRIBUTION_ERROR
|
||
new Guid("f0ad095e-8e9c-4bfb-855e-11fb5dd58cea"), // NAÏVE_REALISM
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Illusion_of_asymmetric_insight",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias TELESCOPING_EFFECT = new()
|
||
{
|
||
Id = new Guid("88b90cfb-93f5-429b-b00f-fabe7ada485c"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Telescoping Effect",
|
||
Description =
|
||
"""
|
||
# Telescoping Effect
|
||
In cognitive psychology, the telescoping effect (or telescoping bias) refers to the temporal displacement of an
|
||
event whereby people perceive recent events as being more remote than they are and distant events as being more
|
||
recent than they are. The former is known as backward telescoping or time expansion, and the latter as is known
|
||
as forward telescoping.
|
||
|
||
The approximate time frame in which events switch from being displaced backward in time to forward in time is three
|
||
years, with events occurring three years in the past being equally likely to be reported with forward telescoping
|
||
bias as with backward telescoping bias. Although telescoping occurs in both the forward and backward directions,
|
||
in general the effect is to increase the number of events reported too recently. This net effect in the forward
|
||
direction is because forces that impair memory, such as lack of salience, also impair time perception.
|
||
|
||
Telescoping leads to an over-reporting of the frequency of events. This over-reporting is because participants
|
||
include events beyond the period, either events that are too recent for the target time period (backward
|
||
telescoping) or events that are too old for the target time period (forward telescoping).
|
||
|
||
## Real-world example
|
||
A real-world example of the telescoping effect is the case of Ferdi Elsas, an infamous kidnapper and murderer
|
||
in the Netherlands. When he was let out of prison, most of the general population did not believe he had been
|
||
in prison long enough. Due to forward telescoping, people thought Ferdi Elsas' sentence started more recently
|
||
than it actually did. Telescoping has important real world applications, especially in survey research. Marketing
|
||
firms often use surveys to ask when consumers last bought a product, and government agencies often use surveys
|
||
to discover information about drug abuse or about victimology. Telescoping may bias answers to these questions.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Telescoping_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias ROSY_RETROSPECTION = new()
|
||
{
|
||
Id = new Guid("5e08ec28-0814-499f-82bd-eb7afb2080aa"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Rosy Retrospection",
|
||
Description =
|
||
"""
|
||
# Rosy Retrospection
|
||
Rosy retrospection is a proposed psychological phenomenon of recalling the past more positively than it
|
||
was actually experienced. The highly unreliable nature of human memory is well documented and accepted
|
||
amongst psychologists. Some research suggests a 'blue retrospective' which also exaggerates negative
|
||
emotions.
|
||
|
||
Though it is a cognitive bias which distorts one's view of reality, it is suggested that rosy retrospection
|
||
serves a useful purpose in increasing self-esteem and sense of well-being. Simplifications and exaggerations
|
||
of memories such as occur in rosy retrospection may make it easier for the brain to store long-term memories,
|
||
as removing details may reduce the burden of those memories by requiring the generation and maintenance of
|
||
fewer neural connections.
|
||
|
||
Declinism, the predisposition to view the past more favourably and the future more negatively, may be related
|
||
to cognitive biases like rosy retrospection. Rosy retrospection is very closely related to the concept of
|
||
nostalgia, though the broader phenomenon of nostalgia is not usually seen as based on a biased perspective.
|
||
|
||
The English idiom "rose-colored glasses" or "rose-tinted glasses" refers to perceiving something more
|
||
positively than it is in reality. The Romans occasionally referred to this phenomenon with the Latin phrase
|
||
"memoria praeteritorum bonorum", which translates into English roughly as "memory of good past", or more
|
||
idiomatically as "good old days".
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("b9edd2f0-8503-4eb5-a4c3-369fcb318894"), // HINDSIGHT_BIAS
|
||
new Guid("7bf44f8f-a4b0-404c-8f15-8ca6e3322d32"), // OPTIMISM_BIAS
|
||
new Guid("b57a862b-b490-4d61-96b8-29d548c2eee4"), // POSITIVITY_EFFECT
|
||
new Guid("23e4b2ad-c915-4d47-ab2d-79a3dce2a7e5"), // DECLINISM
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Rosy_retrospection",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias PROJECTION_BIAS = new()
|
||
{
|
||
Id = new Guid("61ca5b76-66d0-4ce2-b260-7fd42696000a"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Projection Bias",
|
||
Description =
|
||
"""
|
||
# Projection Bias
|
||
|
||
## Affective forecasting
|
||
Affective forecasting, also known as hedonic forecasting or the hedonic forecasting mechanism, is the
|
||
prediction of one's affect (emotional state) in the future. As a process that influences preferences,
|
||
decisions, and behavior, affective forecasting is studied by both psychologists and economists, with
|
||
broad applications.
|
||
|
||
## Bias
|
||
Projection bias is the tendency to falsely project current preferences onto a future event. When people
|
||
are trying to estimate their emotional state in the future they attempt to give an unbiased estimate.
|
||
However, people's assessments are contaminated by their current emotional state. Thus, it may be difficult
|
||
for them to predict their emotional state in the future, an occurrence known as mental contamination. For
|
||
example, if a college student was currently in a negative mood because he just found out he failed a test,
|
||
and if the college student forecasted how much he would enjoy a party two weeks later, his current negative
|
||
mood may influence his forecast. In order to make an accurate forecast the student would need to be aware
|
||
that his forecast is biased due to mental contamination, be motivated to correct the bias, and be able to
|
||
correct the bias in the right direction and magnitude.
|
||
|
||
Projection bias can arise from empathy gaps (or hot/cold empathy gaps), which occur when the present and
|
||
future phases of affective forecasting are characterized by different states of physiological arousal,
|
||
which the forecaster fails to take into account. For example, forecasters in a state of hunger are likely
|
||
to overestimate how much they will want to eat later, overlooking the effect of their hunger on future
|
||
preferences. As with projection bias, economists use the visceral motivations that produce empathy gaps
|
||
to help explain impulsive or self-destructive behaviors, such as smoking.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("8e0f2242-6ad8-4e1e-a9e5-b55a4166781a"), // IMPACT_BIAS
|
||
new Guid("e4e091cf-fed3-4c09-9c21-509db0b2729b"), // HOT_COLD_EMPATHY_GAP
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Affective_forecasting#Projection_bias",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias IMPACT_BIAS = new()
|
||
{
|
||
Id = new Guid("8e0f2242-6ad8-4e1e-a9e5-b55a4166781a"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Impact Bias",
|
||
Description =
|
||
"""
|
||
# Impact Bias
|
||
|
||
## Affective forecasting
|
||
Affective forecasting, also known as hedonic forecasting or the hedonic forecasting mechanism, is the
|
||
prediction of one's affect (emotional state) in the future. As a process that influences preferences,
|
||
decisions, and behavior, affective forecasting is studied by both psychologists and economists, with
|
||
broad applications.
|
||
|
||
## Bias
|
||
One of the most common sources of error in affective forecasting across various populations and situations
|
||
is impact bias, the tendency to overestimate the emotional impact of a future event, whether in terms of
|
||
intensity or duration. The tendencies to overestimate intensity and duration are both robust and reliable
|
||
errors found in affective forecasting.
|
||
|
||
One study documenting impact bias examined college students participating in a housing lottery. These students
|
||
predicted how happy or unhappy they would be one year after being assigned to either a desirable or an undesirable
|
||
dormitory. These college students predicted that the lottery outcomes would lead to meaningful differences in
|
||
their own level of happiness, but follow-up questionnaires revealed that students assigned to desirable or
|
||
undesirable dormitories reported nearly the same levels of happiness. Thus, differences in forecasts
|
||
overestimated the impact of the housing assignment on future happiness.
|
||
|
||
Some studies specifically address "durability bias," the tendency to overestimate the length of time future
|
||
emotional responses will last. Even if people accurately estimate the intensity of their future emotions, they
|
||
may not be able to estimate their duration. Durability bias is generally stronger in reaction to negative events.
|
||
This is important because people tend to work toward events they believe will cause lasting happiness, and according
|
||
to durability bias, people might be working toward the wrong things. Similar to impact bias, durability bias causes
|
||
a person to overemphasize where the root cause of their happiness lies.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("61ca5b76-66d0-4ce2-b260-7fd42696000a"), // PROJECTION_BIAS
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Affective_forecasting#Impact_bias",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias PRO_INNOVATION_BIAS = new()
|
||
{
|
||
Id = new Guid("fa033e14-41f3-45a9-887f-17e30f24c4e5"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Pro-Innovation Bias",
|
||
Description =
|
||
"""
|
||
# Pro-Innovation Bias
|
||
In diffusion of innovation theory, a pro-innovation bias is a belief that innovation should be adopted by the whole
|
||
society without the need for its alteration. The innovation's "champion" has a such strong bias in favor of the
|
||
innovation, that they may not see its limitations or weaknesses and continue to promote it nonetheless.
|
||
|
||
Example:
|
||
A feeling of nuclear optimism emerged in the 1950s in which it was believed that all power generators in the future
|
||
would be atomic in nature. The atomic bomb would render all conventional explosives obsolete and nuclear power plants
|
||
would do the same for power sources such as coal and oil. There was a general feeling that everything would use a
|
||
nuclear power source of some sort, in a positive and productive way, from irradiating food to preserve it, to the
|
||
development of nuclear medicine. There would be an age of peace and plenty in which atomic energy would "provide the
|
||
power needed to desalinate water for the thirsty, irrigate the deserts for the hungry, and fuel interstellar travel
|
||
deep into outer space". This use would render the Atomic Age as significant a step in technological progress as the
|
||
first smelting of Bronze, of Iron, or the commencement of the Industrial Revolution.
|
||
|
||
Roger Smith, then chairman of General Motors, said in 1986: "By the turn of the century, we will live in a paperless
|
||
society." In the late 20th century, there were many predictions of this kind. This transformation has so far not taken
|
||
place.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Pro-innovation_bias",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias TIME_SAVING_BIAS = new()
|
||
{
|
||
Id = new Guid("f262db5e-b668-4bf9-9591-e38e153717da"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Time-Saving Bias",
|
||
Description =
|
||
"""
|
||
# Time-Saving Bias
|
||
Time-saving bias is a concept that describes people's tendency to misestimate the time that could be saved (or
|
||
lost) when increasing (or decreasing) speed. In general, people underestimate the time that could be saved when
|
||
increasing from a relatively low speed—e.g., 25 mph (40 km/h) or 40 mph (64 km/h)—and overestimate the time that
|
||
could be saved when increasing from a relatively high speed—e.g., 55 mph (89 km/h) or 90 mph (140 km/h). People
|
||
also underestimate the time that could be lost when decreasing from a low speed and overestimate the time that
|
||
could be lost when decreasing from a high speed.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Time-saving_bias",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias PLANNING_FALLACY = new()
|
||
{
|
||
Id = new Guid("144a4177-96fa-428f-8f42-bd7c3671c8a6"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Planning Fallacy",
|
||
Description =
|
||
"""
|
||
# Planning Fallacy
|
||
The planning fallacy is a phenomenon in which predictions about how much time will be needed to complete a future task
|
||
display an optimism bias and underestimate the time needed. This phenomenon sometimes occurs regardless of the individual's
|
||
knowledge that past tasks of a similar nature have taken longer to complete than generally planned. The bias affects
|
||
predictions only about one's own tasks. On the other hand, when outside observers predict task completion times, they
|
||
tend to exhibit a pessimistic bias, overestimating the time needed. The planning fallacy involves estimates of task
|
||
completion times more optimistic than those encountered in similar projects in the past.
|
||
|
||
Real-world examples:
|
||
|
||
- The Sydney Opera House was expected to be completed in 1963. A scaled-down version opened in 1973, a decade later.
|
||
The original cost was estimated at $7 million, but its delayed completion led to a cost of $102 million.
|
||
|
||
- The Eurofighter Typhoon defense project took six years longer than expected, with an overrun cost of 8 billion euros.
|
||
|
||
- The Big Dig which undergrounded the Boston Central Artery was completed seven years later than planned, for $8.08
|
||
billion on a budget of $2.8 billion (in 1988 dollars).
|
||
|
||
- The Denver International Airport opened sixteen months later than scheduled, with a total cost of $4.8 billion,
|
||
over $2 billion more than expected.
|
||
|
||
- The Berlin Brandenburg Airport is another case. After 15 years of planning, construction began in 2006, with the
|
||
opening planned for October 2011. There were numerous delays. It was finally opened on October 31, 2020. The
|
||
original budget was €2.83 billion; current projections are close to €10.0 billion.
|
||
|
||
- Olkiluoto Nuclear Power Plant Unit 3 faced severe delay and a cost overrun. The construction started in 2005 and
|
||
was expected to be completed by 2009, but completed only in 2023. Initially, the estimated cost of the project was
|
||
around 3 billion euros, but the cost has escalated to approximately 10 billion euros.
|
||
|
||
- California High-Speed Rail is still under construction, with tens of billions of dollars in overruns expected,
|
||
and connections to major cities postponed until after completion of the rural segment.
|
||
|
||
- The James Webb Space Telescope went over budget by approximately 9 billion dollars, and was sent into orbit 14
|
||
years later than its originally planned launch date.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("7bf44f8f-a4b0-404c-8f15-8ca6e3322d32"), // OPTIMISM_BIAS
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Planning_fallacy",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias PESSIMISM_BIAS = new()
|
||
{
|
||
Id = new Guid("67041978-ac8e-4254-ae2c-509e7301619f"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Pessimism Bias",
|
||
Description =
|
||
"""
|
||
# Pessimism Bias
|
||
The opposite of optimism bias is pessimism bias (or pessimistic bias), because the principles of the optimistic
|
||
bias continue to be in effect in situations where individuals regard themselves as worse off than others. Optimism
|
||
may occur from either a distortion of personal estimates, representing personal optimism, or a distortion for others,
|
||
representing personal pessimism.
|
||
|
||
Pessimism bias is an effect in which people exaggerate the likelihood that negative things will happen to them. It
|
||
contrasts with optimism bias. People with depression are particularly likely to exhibit pessimism bias. Surveys of
|
||
smokers have found that their ratings of their risk of heart disease showed a small but significant pessimism bias;
|
||
however, the literature as a whole is inconclusive.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("7bf44f8f-a4b0-404c-8f15-8ca6e3322d32"), // OPTIMISM_BIAS
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Optimism_bias#Pessimism_bias",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias DECLINISM = new()
|
||
{
|
||
Id = new Guid("23e4b2ad-c915-4d47-ab2d-79a3dce2a7e5"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Declinism",
|
||
Description =
|
||
"""
|
||
# Declinism
|
||
Declinism is the belief that a society or institution is tending towards decline. Particularly, it is the
|
||
predisposition, caused by cognitive biases such as rosy retrospection, to view the past more favourably and
|
||
the future more negatively. "The great summit of declinism" according to Adam Gopnick, "was established in
|
||
1918, in the book that gave decline its good name in publishing: the German historian Oswald Spengler's
|
||
best-selling, thousand-page work *The Decline of the West*."
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("5e08ec28-0814-499f-82bd-eb7afb2080aa"), // ROSY_RETROSPECTION
|
||
new Guid("8533edf9-3117-48c5-8f78-efbd996911f0"), // CONSERVATISM_BIAS
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Declinism",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias MORAL_LUCK = new()
|
||
{
|
||
Id = new Guid("7534480a-1abf-40d5-acec-ace1bfc5be3a"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Moral Luck",
|
||
Description =
|
||
"""
|
||
# Moral Luck
|
||
Moral luck describes circumstances whereby a moral agent is assigned moral blame or praise for an action or
|
||
its consequences even if it is clear that said agent did not have full control over either the action or its
|
||
consequences.
|
||
|
||
Example: There are two people driving cars, Driver A and Driver B. They are alike in every way. Driver A is
|
||
driving down a road and in a moment of inattention runs a red light as a child is crossing the street. Driver
|
||
A slams the brakes, swerves, and does everything to try to avoid hitting the child. Alas, the car hits and
|
||
kills the child. Driver B in the meantime also runs a red light, but since no one is crossing, gets a traffic
|
||
ticket but nothing more.
|
||
|
||
If it is given that moral responsibility should only be relevant when the agent voluntarily performed or
|
||
failed to perform some action, Drivers A and B should be blamed equally, or praised equally, as may be
|
||
the case. However, due to the effect of Moral Luck, if a bystander were asked to morally evaluate Drivers
|
||
A and B, there is very good reason to expect them to say that Driver A is due more moral blame than Driver B.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Moral_luck",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias OUTCOME_BIAS = new()
|
||
{
|
||
Id = new Guid("a3f4415d-b7fa-4668-bcc2-20c79f714bdd"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Outcome Bias",
|
||
Description =
|
||
"""
|
||
# Outcome Bias
|
||
The outcome bias is an error made in evaluating the quality of a decision when the outcome of that decision is
|
||
already known. Specifically, the outcome effect occurs when the same "behavior produce[s] more ethical condemnation
|
||
when it happen[s] to produce bad rather than good outcome, even if the outcome is determined by chance."
|
||
|
||
While similar to the hindsight bias, the two phenomena are markedly different. Hindsight bias focuses on memory
|
||
distortion to favor the actor, while the outcome bias focuses exclusively on weighting the outcome heavier than
|
||
other pieces of information in deciding if a past decision was correct.
|
||
|
||
The outcome bias is closely related to the philosophical concept of moral luck as in both concepts, the evaluation
|
||
of actions is influenced by factors that are not logically justifiable.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("7534480a-1abf-40d5-acec-ace1bfc5be3a"), // MORAL_LUCK
|
||
new Guid("b9edd2f0-8503-4eb5-a4c3-369fcb318894"), // HINDSIGHT_BIAS
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Outcome_bias",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias HINDSIGHT_BIAS = new()
|
||
{
|
||
Id = new Guid("b9edd2f0-8503-4eb5-a4c3-369fcb318894"),
|
||
Category = BiasCategory.NOT_ENOUGH_MEANING,
|
||
Name = "Hindsight Bias",
|
||
Description =
|
||
"""
|
||
# Hindsight Bias
|
||
Hindsight bias, also known as the knew-it-all-along phenomenon or creeping determinism, is the common tendency
|
||
for people to perceive past events as having been more predictable than they were. After an event has occurred,
|
||
people often believe that they could have predicted or perhaps even known with a high degree of certainty what
|
||
the outcome of the event would be before it occurred. Hindsight bias may cause distortions of memories of what
|
||
was known or believed before an event occurred and is a significant source of overconfidence in one’s ability
|
||
to predict the outcomes of future events. Examples of hindsight bias can be seen in the writings of historians
|
||
describing the outcomes of battles, in physicians’ recall of clinical trials, and in criminal or civil trials
|
||
as people tend to assign responsibility on the basis of the supposed predictability of accidents.
|
||
|
||
Hindsight bias has both positive and negative consequences. The bias also plays a role in the process of
|
||
decision-making within the medical field.
|
||
|
||
Positive consequences of hindsight bias is an increase in one's confidence and performance, as long as the bias
|
||
distortion is reasonable and does not create overconfidence. Another positive consequence is that one's
|
||
self-assurance of their knowledge and decision-making, even if it ends up being a poor decision, can be
|
||
beneficial to others; allowing others to experience new things or to learn from those who made the poor
|
||
decisions.
|
||
|
||
Negative: Hindsight bias causes overconfidence in one's performance relative to others. Hindsight bias
|
||
decreases one's rational thinking because of when a person experiences strong emotions, which in turn
|
||
decreases rational thinking. Another negative consequence of hindsight bias is the interference of one's
|
||
ability to learn from experience, as a person is unable to look back on past decisions and learn from
|
||
mistakes. A third consequence is a decrease in sensitivity toward a victim by the person who caused the
|
||
wrongdoing. The person demoralizes the victim and does not allow for a correction of behaviors and actions.
|
||
|
||
Medical decision-making: Hindsight bias may lead to overconfidence and malpractice in regards to physicians.
|
||
Hindsight bias and overconfidence is often attributed to the number of years of experience the physician has.
|
||
After a procedure, physicians may have a "knew it the whole time" attitude, when in reality they may not have
|
||
known it. Medical decision support systems are designed to assist physicians in diagnosis and treatment, and
|
||
have been suggested as a way to counteract hindsight bias. However, these decision support systems come with
|
||
drawbacks, as going against a recommended decision resulted in more punitive jury outcomes when physicians
|
||
were found liable for causing harm.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("af63ce77-f6c6-4e0f-8a9e-3daedc497f9a"), // CONFIRMATION_BIAS
|
||
new Guid("697f58a7-45d7-4268-8951-81681fb005de"), // CURSE_OF_KNOWLEDGE
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Hindsight_bias",
|
||
],
|
||
};
|
||
|
||
#endregion
|
||
|
||
#region NEED_TO_ACT_FAST
|
||
|
||
private static readonly Bias LESS_IS_BETTER_EFFECT = new()
|
||
{
|
||
Id = new Guid("593f2a10-46a6-471e-9ab3-86df740df6f2"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Less-Is-Better Effect",
|
||
Description =
|
||
"""
|
||
# Less-Is-Better Effect
|
||
The less-is-better effect is a type of preference reversal that occurs when the lesser
|
||
or smaller alternative of a proposition is preferred when evaluated separately, but not
|
||
evaluated together.
|
||
|
||
In a 1998 study, Hsee, a professor at the Graduate School of Business of The University
|
||
of Chicago, discovered a less-is-better effect in three contexts:
|
||
|
||
- (1) a person giving a $45 scarf (from scarves ranging from $5-$50) as a gift was
|
||
perceived to be more generous than one giving a $55 coat (from coats ranging from $50-$500);
|
||
|
||
- (2) an overfilled ice cream serving with 7 oz of ice cream was valued more than an underfilled
|
||
serving with 8 oz of ice cream;
|
||
|
||
- (3) a dinnerware set with 24 intact pieces was judged more favourably than one with 31 intact
|
||
pieces (including the same 24) plus a few broken ones.
|
||
|
||
Hsee noted that the less-is-better effect was observed "only when the options were evaluated
|
||
separately, and reversed itself when the options were juxtaposed.” Hsee explained these seemingly
|
||
counterintuitive results “in terms of the evaluability hypothesis, which states that separate
|
||
evaluations of objects are often influenced by attributes that are easy to evaluate rather than
|
||
by those that are important."
|
||
|
||
The less-is-better effect occurs only under specific circumstances. Evidence has shown that it
|
||
manifests itself only when the options are evaluated individually; it disappears when they are
|
||
assessed jointly. "If the options are put right next to each other, the effect disappears, as
|
||
people see the true value of both," states one source. "It's just the gifts in isolation that
|
||
give people a flipped sense of happiness and gratitude."
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Less-is-better_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias OCCAMS_RAZOR = new()
|
||
{
|
||
Id = new Guid("3d5e3115-a98e-4d11-9760-4a3ddbbe6c69"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Occam's Razor",
|
||
Description =
|
||
"""
|
||
# Occam’s Razor
|
||
Among competing hypotheses, the one with the fewest assumptions should be selected. Alternatively,
|
||
other things being equal, simpler explanations are generally better than more complex ones.
|
||
Controversial. This is not a cognitive bias. It is a heuristic, but not one that deviates from
|
||
rationality in judgment.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Occam%27s_razor",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias CONJUNCTION_FALLACY = new()
|
||
{
|
||
Id = new Guid("b0c60f50-cc40-4bde-996c-1833741622a0"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Conjunction Fallacy",
|
||
Description =
|
||
"""
|
||
# Conjunction Fallacy
|
||
The conjunction fallacy (also known as the Linda problem) is an inference that a conjoint set of two or more
|
||
specific conclusions is likelier than any single member of that same set, in violation of the laws of
|
||
probability. It is a type of formal fallacy.
|
||
|
||
The most often-cited example of this fallacy originated with Amos Tversky and Daniel Kahneman:
|
||
|
||
"Linda is 31 years old, single, outspoken, and very bright. She majored in philosophy. As a student, she
|
||
was deeply concerned with issues of discrimination and social justice, and also participated in anti-nuclear
|
||
demonstrations."
|
||
|
||
Which is more probable?
|
||
|
||
- Linda is a bank teller.
|
||
- Linda is a bank teller and is active in the feminist movement.
|
||
|
||
The majority of those asked chose option 2. However, the probability of two events occurring together
|
||
(that is, in conjunction) is always less than or equal to the probability of either one occurring itself.
|
||
|
||
Tversky and Kahneman argue that most people get this problem wrong because they use a heuristic (an easily
|
||
calculated) procedure called representativeness to make this kind of judgment: Option 2 seems more
|
||
"representative" of Linda from the description of her, even though it is clearly mathematically less likely.
|
||
|
||
## Debiasing
|
||
Drawing attention to set relationships, using frequencies instead of probabilities, and/or thinking
|
||
diagrammatically (e.g. use a Venn diagram) sharply reduce the error in some forms of the conjunction
|
||
fallacy.
|
||
|
||
In one experiment the question of the Linda problem was reformulated as follows:
|
||
|
||
"There are 100 persons who fit the description above (that is, Linda's). How many of them are:
|
||
|
||
- Bank tellers? __ of 100
|
||
- Bank tellers and active in the feminist movement? __ of 100"
|
||
|
||
Whereas previously 85% of participants gave the wrong answer (bank teller and active in the feminist
|
||
movement), in experiments done with this questioning the proportion of incorrect answers is dramatically
|
||
reduced (to ~20%). Participants were forced to use a mathematical approach and thus recognized the
|
||
difference more easily.
|
||
|
||
However, in some tasks only based on frequencies, not on stories, that used clear logical formulations,
|
||
conjunction fallacies continued to occur dominantly, with only few exceptions, when the observed pattern
|
||
of frequencies resembled a conjunction.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Conjunction_fallacy",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias DELMORE_EFFECT = new()
|
||
{
|
||
Id = new Guid("93a3d088-183f-47e7-a010-721f1cd6bac8"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Delmore Effect",
|
||
Description =
|
||
"""
|
||
# Delmore Effect
|
||
The Delmore effect is about how we tend to set clearer and more detailed goals for less important areas
|
||
of our lives. In other words, we distract ourselves from the most important tasks by focusing on the
|
||
easy stuff instead.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://www.42courses.com/blog/home/2022/9/2/42-effects-you-should-know-part-2",
|
||
"https://bias.transhumanity.net/delmore-effect/",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias PARKINSONS_LAW_OF_TRIVIALITY = new()
|
||
{
|
||
Id = new Guid("d3ec6a5d-91cf-4aec-8541-bd87e1ad834b"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Law of Triviality",
|
||
Description =
|
||
"""
|
||
# Law of Triviality
|
||
The law of triviality is C. Northcote Parkinson's 1957 argument that people within an organization commonly
|
||
give disproportionate weight to trivial issues. Parkinson provides the example of a fictional committee
|
||
whose job was to approve the plans for a nuclear power plant spending the majority of its time on discussions
|
||
about relatively minor but easy-to-grasp issues, such as what materials to use for the staff bicycle shed,
|
||
while neglecting the proposed design of the plant itself, which is far more important and a far more difficult
|
||
and complex task.
|
||
|
||
The law has been applied to software development and other activities. The terms bicycle-shed effect,
|
||
bike-shed effect, and bike-shedding were coined based on Parkinson's example; it was popularized in the
|
||
Berkeley Software Distribution community by the Danish software developer Poul-Henning Kamp in 1999
|
||
and, due to that, has since become popular within the field of software development generally.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("b9c06da1-d2eb-4871-8159-a2a6d25e9eff"), // DUNNING_KRUGER_EFFECT
|
||
new Guid("ad32d669-fc79-44c9-a570-609e1ccdc799"), // OMISSION_BIAS
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Law_of_triviality",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias RHYME_AS_REASON_EFFECT = new()
|
||
{
|
||
Id = new Guid("0d290221-81a0-4e44-bdec-30709117d90d"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Rhyme as Reason Effect",
|
||
Description =
|
||
"""
|
||
# Rhyme as Reason Effect
|
||
The rhyme-as-reason effect, also known as the Eaton–Rosen phenomenon, is a cognitive bias where sayings
|
||
or aphorisms are perceived as more accurate or truthful when they rhyme. In experiments, participants
|
||
evaluated variations of sayings that either rhymed or did not rhyme. Those that rhymed were consistently
|
||
judged as more truthful, even when the meaning was controlled for. For instance, the rhyming saying "What
|
||
sobriety conceals, alcohol reveals" was rated as more accurate on average than its non-rhyming counterpart,
|
||
"What sobriety conceals, alcohol unmasks," across different groups of subjects (each group assessed the
|
||
accuracy of only one version of the statement).
|
||
|
||
This effect may be explained by the Keats heuristic, which suggests that people assess a statement's truth
|
||
based on its aesthetic qualities. Another explanation is the fluency heuristic, which posits that statements
|
||
are preferred due to their ease of cognitive processing.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("cadafb8f-d1ed-4c92-9c29-2f1cb0797a66"), // ILLUSORY_TRUTH_EFFECT
|
||
new Guid("1c5aa90a-e732-4f45-bf26-1b86c49a82f9"), // BELIEF_BIAS
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Rhyme-as-reason_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias BELIEF_BIAS = new()
|
||
{
|
||
Id = new Guid("1c5aa90a-e732-4f45-bf26-1b86c49a82f9"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Belief Bias",
|
||
Description =
|
||
"""
|
||
# Belief Bias
|
||
Belief bias is the tendency to judge the strength of arguments based on the plausibility of their conclusion
|
||
rather than how strongly they justify that conclusion. A person is more likely to accept an argument that
|
||
supports a conclusion that aligns with their values, beliefs and prior knowledge, while rejecting counter
|
||
arguments to the conclusion. Belief bias is an extremely common and therefore significant form of error;
|
||
we can easily be blinded by our beliefs and reach the wrong conclusion. Belief bias has been found to
|
||
influence various reasoning tasks, including conditional reasoning, relation reasoning and transitive reasoning.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Belief_bias",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias INFORMATION_BIAS = new()
|
||
{
|
||
Id = new Guid("d0e251bb-3e09-43f5-8c5e-bc933e743509"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Information Bias",
|
||
Description =
|
||
"""
|
||
# Information Bias
|
||
The tendency to seek information when it does not affect action. An example of information bias is believing
|
||
that the more information that can be acquired to make a decision, the better, even if that extra information
|
||
is irrelevant for the decision.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Information_bias_(psychology)",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias AMBIGUITY_EFFECT = new()
|
||
{
|
||
Id = new Guid("e9b00144-0cb3-46de-8a68-09daa00de1e4"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Ambiguity Effect",
|
||
Description =
|
||
"""
|
||
# Ambiguity Effect
|
||
The ambiguity effect is a cognitive tendency where decision making is affected by a lack of information, or
|
||
"ambiguity". The effect implies that people tend to select options for which the probability of a favorable
|
||
outcome is known, over an option for which the probability of a favorable outcome is unknown. The effect was
|
||
first described by Daniel Ellsberg in 1961.
|
||
|
||
One possible explanation of the effect is that people have a rule of thumb (heuristic) to avoid options where
|
||
information is missing. This will often lead them to seek out the missing information. In many cases, though,
|
||
the information cannot be obtained. The effect is often the result of calling some particular missing piece of
|
||
information to the person's attention.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("44c6efd7-53f1-4d22-82fe-25e941390089"), // NEGLECT_OF_PROBABILITY
|
||
new Guid("73ca0caa-25e5-4edb-91d4-f375a773f82c"), // APPEAL_TO_POSSIBILITY
|
||
new Guid("b0c60f50-cc40-4bde-996c-1833741622a0"), // CONJUNCTION_FALLACY
|
||
new Guid("656c78c9-d75a-4c07-a80d-f3a5026f859c"), // PSEUDOCERTAINTY_EFFECT
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Ambiguity_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias STATUS_QUO_BIAS = new()
|
||
{
|
||
Id = new Guid("b9e05a25-ac09-407d-8aee-f54a04decf0b"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Status Quo Bias",
|
||
Description =
|
||
"""
|
||
# Status Quo Bias
|
||
A status quo bias or default bias is a cognitive bias which results from a preference for the maintenance
|
||
of one's existing state of affairs. The current baseline (or status quo) is taken as a reference point,
|
||
and any change from that baseline is perceived as a loss or gain. Corresponding to different alternatives,
|
||
this current baseline or default option is perceived and evaluated by individuals as a positive.
|
||
|
||
Status quo bias should be distinguished from a rational preference for the status quo ante, as when the
|
||
current state of affairs is objectively superior to the available alternatives, or when imperfect information
|
||
is a significant problem. A large body of evidence, however, shows that status quo bias frequently affects
|
||
human decision-making. Status quo bias should also be distinguished from psychological inertia, which refers
|
||
to a lack of intervention in the current course of affairs.
|
||
|
||
The bias intersects with other non-rational cognitive processes such as loss aversion, in which losses
|
||
comparative to gains are weighed to a greater extent. Further non-rational cognitive processes include
|
||
existence bias, endowment effect, longevity, mere exposure, and regret avoidance. Experimental evidence
|
||
for the detection of status quo bias is seen through the use of the reversal test. A vast amount of
|
||
experimental and field examples exist. Behaviour in regard to economics, retirement plans, health, and
|
||
ethical choices show evidence of the status quo bias.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("b81482f8-b2cf-4b86-a5a4-fcd29aee4e69"), // ENDOWMENT_EFFECT
|
||
new Guid("ad32d669-fc79-44c9-a570-609e1ccdc799"), // OMISSION_BIAS
|
||
new Guid("ad3ed908-c56e-411b-a130-8af8574ff67b"), // LOSS_AVERSION
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Status_quo_bias",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias SOCIAL_COMPARISON_BIAS = new()
|
||
{
|
||
Id = new Guid("09527928-6417-4eea-9719-d8ed4748691f"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Social Comparison Bias",
|
||
Description =
|
||
"""
|
||
# Social Comparison Bias
|
||
Social comparison bias is the tendency to have feelings of dislike and competitiveness with someone seen as
|
||
physically, socially, or mentally better than oneself. Social comparison bias or social comparison theory is
|
||
the idea that individuals determine their own worth based on how they compare to others. The theory was
|
||
developed in 1954 by psychologist Leon Festinger. This can be compared to social comparison, which is
|
||
believed to be central to achievement motivation, feelings of injustice, depression, jealousy, and people's
|
||
willingness to remain in relationships or jobs. The basis of the theory is that people are believed to
|
||
compete for the best outcome in relation to their peers. For example, one might make a comparison between the
|
||
low-end department stores they go to frequently and the designer stores of their peers. Such comparisons may
|
||
evoke feelings of resentment, anger, and envy with their peers. This bias revolves mostly around wealth and
|
||
social status; it is unconscious and people who make these are largely unaware of them. In most cases, people
|
||
try to compare themselves to those in their peer group or with whom they are similar.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Social_comparison_bias",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias DECOY_EFFECT = new()
|
||
{
|
||
Id = new Guid("c8a532e9-5958-4894-aa0d-29ed6412780f"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Decoy Effect",
|
||
Description =
|
||
"""
|
||
# Decoy Effect
|
||
In marketing, the decoy effect (or attraction effect or asymmetric dominance effect) is the phenomenon
|
||
whereby consumers will tend to have a specific change in preference between two options when also presented
|
||
with a third option that is asymmetrically dominated. An option is asymmetrically dominated when it is
|
||
inferior in all respects to one option; but, in comparison to the other option, it is inferior in some
|
||
respects and superior in others. In other words, in terms of specific attributes determining preferences,
|
||
it is completely dominated by (i.e., inferior to) one option and only partially dominated by the other.
|
||
When the asymmetrically dominated option is present, a higher percentage of consumers will prefer the
|
||
dominating option than when the asymmetrically dominated option is absent. The asymmetrically dominated
|
||
option is therefore a decoy serving to increase preference for the dominating option. The decoy effect
|
||
is also an example of the violation of the independence of irrelevant alternatives axiom of decision
|
||
theory. More simply, when deciding between two options, an unattractive third option can change the
|
||
perceived preference between the other two.
|
||
|
||
The decoy effect is considered particularly important in choice theory because it is a violation of the
|
||
assumption of "regularity" present in all axiomatic choice models, for example in a Luce model of choice.
|
||
Regularity means that it should not be possible for the market share of any alternative to increase when
|
||
another alternative is added to the choice set. The new alternative should reduce, or at best leave unchanged,
|
||
the choice share of existing alternatives. Regularity is violated in the example shown below where a new
|
||
alternative C not only changes the relative shares of A and B but actually increases the share of A in
|
||
absolute terms. Similarly, the introduction of a new alternative D increases the share of B in absolute
|
||
terms.
|
||
|
||
## Example
|
||
Suppose there is a consideration set (options to choose from in a menu) that involves smartphones. Consumers
|
||
will generally see higher storage capacity (number of GB) and lower price as positive attributes; while some
|
||
consumers may want a device that can store more photos, music, etc., other consumers will want a device that
|
||
costs less. In Consideration Set 1, two devices are available:
|
||
|
||
Consideration Set 1:
|
||
- A: $400, 300GB
|
||
- B: $300, 200GB
|
||
|
||
In this case, some consumers will prefer A for its greater storage capacity, while others will prefer B for
|
||
its lower price.
|
||
|
||
Now suppose that a new player, C, the "decoy", is added to the market; it is more expensive than both A, the
|
||
"target", and B, the "competitor", and has more storage than B but less than A:
|
||
|
||
Consideration Set 2:
|
||
- A (target): $400, 300GB
|
||
- B (competitor): $300, 200GB
|
||
- C (decoy): $450, 250GB
|
||
|
||
The addition of decoy C — which consumers would presumably avoid, given that a lower price can be paid for a
|
||
model with more storage—causes A, the dominating option, to be chosen more often than if only the two choices
|
||
in Consideration Set 1 existed; C affects consumer preferences by acting as a basis of comparison for A and B.
|
||
Because A is better than C in both respects, while B is only partially better than C, more consumers will
|
||
prefer A now than did before. C is therefore a decoy whose sole purpose is to increase sales of A.
|
||
|
||
Conversely, suppose that instead of C, a player D is introduced that has less storage than both A and B, and
|
||
that is more expensive than B but not as expensive as A:
|
||
|
||
Consideration Set 3:
|
||
- A (competitor): $400, 300GB
|
||
- B (target): $300, 200GB
|
||
- D (decoy): $350, 150GB
|
||
|
||
The result here is similar: consumers will not prefer D, because it is not as good as B in any respect. However,
|
||
whereas C increased preference for A, D has the opposite effect, increasing preference for B.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Decoy_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias REACTANCE = new()
|
||
{
|
||
Id = new Guid("d3c2cb4b-ec29-4cf3-a485-9a98e9f1f223"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Reactance",
|
||
Description =
|
||
"""
|
||
# Reactance
|
||
In psychology, reactance is an unpleasant motivational reaction to offers, persons, rules, regulations, advice, or
|
||
recommendations that are perceived to threaten or eliminate specific behavioral freedoms. Reactance occurs when an
|
||
individual feels that an agent is attempting to limit one's choice of response and/or range of alternatives.
|
||
|
||
Reactance can occur when someone is heavily pressured into accepting a certain view or attitude. Reactance can
|
||
encourage an individual to adopt or strengthen a view or attitude which is indeed contrary to that which was
|
||
intended — which is to say, to a response of noncompliance — and can also increase resistance to persuasion.
|
||
Some individuals might employ reverse psychology in a bid to exploit reactance for their benefit, in an attempt
|
||
to influence someone to choose the opposite of what is being requested. Reactance can occur when an individual
|
||
senses that someone is trying to compel them to do something; often the individual will offer resistance and
|
||
attempt to extricate themselves from the situation.
|
||
|
||
Some individuals are naturally high in reactance, a personality characteristic called trait reactance.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("a9c7faa7-2368-4be5-9eda-a37ffd8f7ab1"), // REVERSE_PSYCHOLOGY
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Reactance_(psychology)",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias REVERSE_PSYCHOLOGY = new()
|
||
{
|
||
Id = new Guid("a9c7faa7-2368-4be5-9eda-a37ffd8f7ab1"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Reverse Psychology",
|
||
Description =
|
||
"""
|
||
# Reverse Psychology
|
||
Reverse psychology is a technique involving the assertion of a belief or behavior that is opposite to the one desired, with the expectation
|
||
that this approach will encourage the subject of the persuasion to do what is actually desired. This technique relies on the psychological
|
||
phenomenon of reactance, in which a person has a negative emotional reaction to being persuaded, and thus chooses the option which is being
|
||
advocated against. This may work especially well on a person who is resistant by nature, while direct requests work best for people who are
|
||
compliant. The one being manipulated is usually unaware of what is really going on.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("d3c2cb4b-ec29-4cf3-a485-9a98e9f1f223"), // REACTANCE
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Reverse_psychology",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias SYSTEM_JUSTIFICATION = new()
|
||
{
|
||
Id = new Guid("755c8f9e-b172-4ff7-9797-9cc130bf4939"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "System Justification",
|
||
Description =
|
||
"""
|
||
# System Justification
|
||
System justification theory is a theory within social psychology that system-justifying beliefs serve a psychologically
|
||
palliative function. It proposes that people have several underlying needs, which vary from individual to individual,
|
||
that can be satisfied by the defense and justification of the status quo, even when the system may be disadvantageous
|
||
to certain people. People have epistemic, existential, and relational needs that are met by and manifest as ideological
|
||
support for the prevailing structure of social, economic, and political norms. Need for order and stability, and thus
|
||
resistance to change or alternatives, for example, can be a motivator for individuals to see the status quo as good,
|
||
legitimate, and even desirable.
|
||
|
||
According to system justification theory, people desire not only to hold favorable attitudes about themselves
|
||
(ego-justification) and the groups to which they belong (group-justification), but also to hold positive attitudes
|
||
about the overarching social structure in which they are entwined and find themselves obligated to (system-justification).
|
||
This system-justifying motive sometimes produces the phenomenon known as out-group favoritism, an acceptance of inferiority
|
||
among low-status groups and a positive image of relatively higher status groups. Thus, the notion that individuals are
|
||
simultaneously supporters and victims of the system-instilled norms is a central idea in system justification theory.
|
||
Additionally, the passive ease of supporting the current structure, when compared to the potential price (material,
|
||
social, psychological) of acting out against the status quo, leads to a shared environment in which the existing social,
|
||
economic, and political arrangements tend to be preferred. Alternatives to the status quo tend to be disparaged, and
|
||
inequality tends to perpetuate.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("b9e05a25-ac09-407d-8aee-f54a04decf0b"), // STATUS_QUO_BIAS
|
||
new Guid("b1cc861b-f445-450b-9bdf-e9d222abdb4e"), // IN_GROUP_FAVORITISM
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/System_justification",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias BELIEF_PERSEVERANCE = new()
|
||
{
|
||
Id = new Guid("bf8f304d-2e8e-4a90-a9c5-7bd56f6058a6"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Belief Perseverance",
|
||
Description =
|
||
"""
|
||
# Belief Perseverance
|
||
Belief perseverance (also known as conceptual conservatism) is maintaining a belief despite new information that
|
||
firmly contradicts it. Since rationality involves conceptual flexibility, belief perseverance is consistent with
|
||
the view that human beings act at times in an irrational manner. Philosopher F.C.S. Schiller holds that belief
|
||
perseverance "deserves to rank among the fundamental 'laws' of nature".
|
||
|
||
If beliefs are strengthened after others attempt to present evidence debunking them, this is known as a backfire
|
||
effect. There are psychological mechanisms by which backfire effects could potentially occur, but the evidence on
|
||
this topic is mixed, and backfire effects are very rare in practice. A 2020 review of the scientific literature on
|
||
backfire effects found that there have been widespread failures to replicate their existence, even under conditions
|
||
that would be theoretically favorable to observing them. Due to the lack of reproducibility, as of 2020 most
|
||
researchers believe that backfire effects are either unlikely to occur on the broader population level, or they
|
||
only occur in very specific circumstances, or they do not exist. For most people, corrections and fact-checking
|
||
are very unlikely to have a negative impact, and there is no specific group of people in which backfire effects
|
||
have been consistently observed.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Belief_perseverance",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias ENDOWMENT_EFFECT = new()
|
||
{
|
||
Id = new Guid("b81482f8-b2cf-4b86-a5a4-fcd29aee4e69"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Endowment Effect",
|
||
Description =
|
||
"""
|
||
# Endowment Effect
|
||
In psychology and behavioral economics, the endowment effect, also known as divestiture aversion, is the finding
|
||
that people are more likely to retain an object they own than acquire that same object when they do not own it.
|
||
The endowment theory can be defined as "an application of prospect theory positing that loss aversion associated
|
||
with ownership explains observed exchange asymmetries."
|
||
|
||
This is typically illustrated in two ways. In a valuation paradigm, people's maximum willingness to pay (WTP) to
|
||
acquire an object is typically lower than the least amount they are willing to accept (WTA) to give up that same
|
||
object when they own it—even when there is no cause for attachment, or even if the item was only obtained minutes
|
||
ago. In an exchange paradigm, people given a good are reluctant to trade it for another good of similar value.
|
||
For example, participants first given a pen of equal expected value to that of a coffee mug were generally unwilling
|
||
to trade, whilst participants first given the coffee mug were also unwilling to trade it for the pen.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("ad3ed908-c56e-411b-a130-8af8574ff67b"), // LOSS_AVERSION
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Endowment_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias PROCESSING_DIFFICULTY_EFFECT = new()
|
||
{
|
||
Id = new Guid("4f61b9fa-146a-4b6e-b075-f0ba2ee0d9d0"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Processing Difficulty Effect",
|
||
Description =
|
||
"""
|
||
# Processing Difficulty Effect
|
||
That information that takes longer to read and is thought about more (processed with more difficulty) is more easily remembered.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("a4027640-1f52-4ff1-ae13-bd14a30d5b8d"), // LEVELS_OF_PROCESSING_EFFECT
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/List_of_cognitive_biases#Other_memory_biases",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias PSEUDOCERTAINTY_EFFECT = new()
|
||
{
|
||
Id = new Guid("656c78c9-d75a-4c07-a80d-f3a5026f859c"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Pseudocertainty Effect",
|
||
Description =
|
||
"""
|
||
# Pseudocertainty Effect
|
||
In prospect theory, the pseudocertainty effect is the tendency for people to perceive an outcome as certain while it is
|
||
actually uncertain in multi-stage decision making. The evaluation of the certainty of the outcome in a previous stage of
|
||
decisions is disregarded when selecting an option in subsequent stages. Not to be confused with certainty effect, the
|
||
pseudocertainty effect was discovered from an attempt at providing a normative use of decision theory for the certainty
|
||
effect by relaxing the cancellation rule.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("ad3ed908-c56e-411b-a130-8af8574ff67b"), // LOSS_AVERSION
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Pseudocertainty_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias CERTAINTY_EFFECT = new()
|
||
{
|
||
Id = new Guid("ac7d745c-d66e-4886-87d7-ddaba349d4e8"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Certainty Effect",
|
||
Description =
|
||
"""
|
||
# Certainty Effect
|
||
The certainty effect is the psychological effect resulting from the reduction of probability from certain to probable
|
||
(Tversky & Kahneman 1986). It is an idea introduced in prospect theory. Normally a reduction in the probability of
|
||
winning a reward (e.g., a reduction from 80% to 20% in the chance of winning a reward) creates a psychological effect
|
||
such as displeasure to individuals, which leads to the perception of loss from the original probability thus favoring
|
||
a risk-averse decision. However, the same reduction results in a larger psychological effect when it is done from
|
||
certainty than from uncertainty.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Certainty_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias DISPOSITION_EFFECT = new()
|
||
{
|
||
Id = new Guid("4ecb0187-b2e2-446f-87e2-1e32f269e497"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Disposition Effect",
|
||
Description =
|
||
"""
|
||
# Disposition Effect
|
||
The disposition effect is an anomaly discovered in behavioral finance. It relates to the tendency of investors to sell
|
||
assets that have increased in value, while keeping assets that have dropped in value. Hersh Shefrin and Meir Statman
|
||
identified and named the effect in their 1985 paper, which found that people dislike losing significantly more than they
|
||
enjoy winning. The disposition effect has been described as one of the foremost vigorous actualities around individual
|
||
investors because investors will hold stocks that have lost value yet sell stocks that have gained value.
|
||
|
||
In 1979, Daniel Kahneman and Amos Tversky traced the cause of the disposition effect to the so-called "prospect theory".
|
||
The prospect theory proposes that when an individual is presented with two equal choices, one having possible gains and
|
||
the other with possible losses, the individual is more likely to opt for the former choice even though both would yield
|
||
the same economic result.
|
||
|
||
The disposition effect can be minimized by means of a mental approach called "hedonic framing". For example, individuals
|
||
can try to force themselves to think of a single large gain as a number of smaller gains, to think of a number of smaller
|
||
losses as a single large loss, to think of the combination of a major gain and a minor loss as a net minor gain, and, in
|
||
the case of a combined major loss and minor gain, to think of the two separately. In a similar manner, investors show a
|
||
reversed disposition effect when they are framed to think of their investment as progress towards a specific investment
|
||
goal rather than a generic investment.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Disposition_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias ZERO_RISK_BIAS = new()
|
||
{
|
||
Id = new Guid("77553998-bfa7-450e-acd9-586a55064302"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Zero-Risk Bias",
|
||
Description =
|
||
"""
|
||
# Zero-Risk Bias
|
||
Zero-risk bias is a tendency to prefer the complete elimination of risk in a sub-part over alternatives with greater
|
||
overall risk reduction. It often manifests in cases where decision makers address problems concerning health, safety,
|
||
and the environment. Its effect on decision making has been observed in surveys presenting hypothetical scenarios.
|
||
|
||
Zero-risk bias is based on the way people feel better if a risk is eliminated instead of being merely mitigated.
|
||
Scientists identified a zero-risk bias in responses to a questionnaire about a hypothetical cleanup scenario involving
|
||
two hazardous sites X and Y, with X causing 8 cases of cancer annually and Y causing 4 cases annually. The respondents
|
||
ranked three cleanup approaches: two options each reduced the total number of cancer cases by 6, while the third reduced
|
||
the number by 5 and eliminated the cases at site Y. While the latter option featured the worst reduction overall, 42% of
|
||
the respondents ranked it better than at least one of the other options. This conclusion resembled one from an earlier
|
||
economics study that found people were willing to pay high costs to eliminate a risk. It has a normative justification
|
||
since once risk is eliminated, people would have less to worry about and such removal of worry also has utility. It is
|
||
also driven by our preference for winning much more than losing as well as the old instead of the new way, all of which
|
||
cloud the way the world is viewed.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Zero-risk_bias",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias UNIT_BIAS = new()
|
||
{
|
||
Id = new Guid("ff43a9e2-7dde-47ca-a3ef-5a9c2d3117c9"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Unit Bias",
|
||
Description =
|
||
"""
|
||
# Unit Bias
|
||
The standard suggested amount of consumption (e.g., food serving size) is perceived to be appropriate, and a person would
|
||
consume it all even if it is too much for this particular person.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/List_of_cognitive_biases#Other",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias IKEA_EFFECT = new()
|
||
{
|
||
Id = new Guid("565616dc-ed84-42af-b9cc-6fa666cc5d66"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "IKEA Effect",
|
||
Description =
|
||
"""
|
||
# IKEA Effect
|
||
The IKEA effect is a cognitive bias in which consumers place a disproportionately high value on products they
|
||
partially created. The name refers to Swedish manufacturer and furniture retailer IKEA, which sells many items
|
||
of furniture that require assembly. A 2011 study found that subjects were willing to pay 63% more for furniture
|
||
they had assembled themselves than for equivalent pre-assembled items.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("b9c06da1-d2eb-4871-8159-a2a6d25e9eff"), // DUNNING_KRUGER_EFFECT
|
||
new Guid("30deb7d6-4019-4fef-9823-8d8126e54f0a"), // ESCALATION_OF_COMMITMENT
|
||
new Guid("ad32d669-fc79-44c9-a570-609e1ccdc799"), // OMISSION_BIAS
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/IKEA_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias LOSS_AVERSION = new()
|
||
{
|
||
Id = new Guid("ad3ed908-c56e-411b-a130-8af8574ff67b"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Loss Aversion",
|
||
Description =
|
||
"""
|
||
# Loss Aversion
|
||
In cognitive science and behavioral economics, loss aversion refers to a cognitive bias in which the same situation
|
||
is perceived as worse if it is framed as a loss, rather than a gain. It should not be confused with risk aversion,
|
||
which describes the rational behavior of valuing an uncertain outcome at less than its expected value.
|
||
|
||
## Application
|
||
In marketing, the use of trial periods and rebates tries to take advantage of the buyer's tendency to value the good
|
||
more after the buyer incorporates it in the status quo. In past behavioral economics studies, users participate up
|
||
until the threat of loss equals any incurred gains. Methods established by Botond Kőszegi and Matthew Rabin in
|
||
experimental economics illustrates the role of expectation, wherein an individual's belief about an outcome can
|
||
create an instance of loss aversion, whether or not a tangible change of state has occurred.
|
||
|
||
Whether a transaction is framed as a loss or as a gain is important to this calculation. The same change in price
|
||
framed differently, for example as a $5 discount or as a $5 surcharge avoided, has a significant effect on
|
||
consumer behavior. Although traditional economists consider this "endowment effect", and all other effects of
|
||
loss aversion, to be completely irrational, it is important to the fields of marketing and behavioral finance.
|
||
Users in behavioral and experimental economics studies decided to cease participation in iterative money-making
|
||
games when the threat of loss was close to the expenditure of effort, even when the user stood to further their
|
||
gains. Loss aversion coupled with myopia has been shown to explain macroeconomic phenomena, such as the equity
|
||
premium puzzle. Loss aversion to kinship is an explanation for aversion to inheritance tax.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("b81482f8-b2cf-4b86-a5a4-fcd29aee4e69"), // ENDOWMENT_EFFECT
|
||
new Guid("ef521fbb-c20b-47c9-87f8-a571a06a03eb"), // NEGATIVITY_BIAS
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Loss_aversion",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias GENERATION_EFFECT = new()
|
||
{
|
||
Id = new Guid("af442ab1-ffc5-404c-9ee8-3497fe6992ec"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Generation Effect",
|
||
Description =
|
||
"""
|
||
# Generation Effect
|
||
The generation effect is a phenomenon whereby information is better remembered if it is generated from one's own
|
||
mind rather than simply read. Researchers have struggled to fully explain why generated information is better
|
||
recalled than read information, as no single explanation has been comprehensive.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Generation_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias ESCALATION_OF_COMMITMENT = new()
|
||
{
|
||
Id = new Guid("30deb7d6-4019-4fef-9823-8d8126e54f0a"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Escalation of Commitment",
|
||
Description =
|
||
"""
|
||
# Escalation of Commitment
|
||
Escalation of commitment is a human behavior pattern in which an individual or group facing increasingly negative
|
||
outcomes from a decision, action, or investment nevertheless continue the behavior instead of altering course.
|
||
The actor maintains behaviors that are irrational, but align with previous decisions and actions.
|
||
|
||
Economists and behavioral scientists use a related term, sunk-cost fallacy, to describe the justification of
|
||
increased investment of money or effort in a decision, based on the cumulative prior investment ("sunk cost")
|
||
despite new evidence suggesting that the future cost of continuing the behavior outweighs the expected benefit.
|
||
|
||
In sociology, irrational escalation of commitment or commitment bias describe similar behaviors. The phenomenon
|
||
and the sentiment underlying them are reflected in such proverbial images as "throwing good money after bad",
|
||
or "In for a penny, in for a pound", or "It's never the wrong time to make the right decision", or "If you find
|
||
yourself in a hole, stop digging."
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("9a2d58f5-bbf1-4b34-8e1b-f9bcd8814f05"), // SUNK_COST_FALLACY
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Escalation_of_commitment",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias SUNK_COST_FALLACY = new()
|
||
{
|
||
Id = new Guid("9a2d58f5-bbf1-4b34-8e1b-f9bcd8814f05"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Sunk Cost Fallacy",
|
||
Description =
|
||
"""
|
||
# Sunk Cost Fallacy
|
||
The Misconception: You make rational decisions based on the future value of objects, investments and experiences.
|
||
The Truth: Your decisions are tainted by the emotional investments you accumulate, and the more you invest in
|
||
something the harder it becomes to abandon it.
|
||
|
||
Example: R&D costs. Once spent, such costs are sunk and should have no effect on future pricing decisions. So a
|
||
pharmaceutical company's attempt to justify high prices because of the need to recoup R&D expenses is fallacious.
|
||
The company will charge market prices whether R&D had cost one dollar or one million dollars. However, R&D costs,
|
||
and the ability to recoup those costs, are a factor in deciding whether to spend the money on R&D. It’s important
|
||
to distinguish that while justifying high prices on past R&D is a fallacy, raising prices in order to finance
|
||
future R&D is not.
|
||
|
||
Counterpoint: It is sometimes not that simple. In a broad range of situations, it is rational for people to condition
|
||
behavior on sunk costs, because of informational content, reputational concerns, or financial and time constraints.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("30deb7d6-4019-4fef-9823-8d8126e54f0a"), // ESCALATION_OF_COMMITMENT
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Sunk_cost#Fallacy_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias IDENTIFIABLE_VICTIM_EFFECT = new()
|
||
{
|
||
Id = new Guid("0c18a8bd-5e5f-4cf0-a90e-47dd7a421035"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Identifiable Victim Effect",
|
||
Description =
|
||
"""
|
||
# Identifiable Victim Effect
|
||
The identifiable victim effect is the tendency of individuals to offer greater aid when a specific, identifiable
|
||
person ("victim") is observed under hardship, as compared to a large, vaguely defined group with the same need.
|
||
|
||
The identifiable victim effect has two components. People are more inclined to help an identified victim than an
|
||
unidentified one, and people are more inclined to help a single identified victim than a group of identified victims.
|
||
Although helping an identified victim may be commendable, the identifiable victim effect is considered a cognitive
|
||
bias. From a consequentialist point of view, the cognitive error is the failure to offer N times as much help to N
|
||
unidentified victims.
|
||
|
||
The identifiable victim effect has a mirror image that is sometimes called the identifiable perpetrator effect.
|
||
Research has shown that individuals are more inclined to mete out punishment, even at their own expense, when they
|
||
are punishing a specific, identified perpetrator.
|
||
|
||
The conceptualization of the identifiable victim effect as it is known today is commonly attributed to American
|
||
economist Thomas Schelling. He wrote that harm to a particular person invokes "anxiety and sentiment, guilt and awe,
|
||
responsibility and religion, [but]…most of this awesomeness disappears when we deal with statistical death".
|
||
|
||
Historical figures from Joseph Stalin to Mother Teresa are credited with statements that epitomize the identifiable
|
||
victim effect. The remark "One death is a tragedy; a million deaths is a statistic" is widely, although probably
|
||
incorrectly, attributed to Stalin. The remark "If I look at the mass I will never act. If I look at the one, I
|
||
will," is attributed to Mother Teresa.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Identifiable_victim_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias APPEAL_TO_NOVELTY = new()
|
||
{
|
||
Id = new Guid("2d57f4d6-e599-4738-812a-c12cef877779"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Appeal to Novelty",
|
||
Description =
|
||
"""
|
||
# Appeal to Novelty
|
||
The appeal to novelty (also called appeal to modernity or argumentum ad novitatem) is a fallacy in which one
|
||
prematurely claims that an idea or proposal is correct or superior, exclusively because it is new and modern.
|
||
In a controversy between status quo and new inventions, an appeal to novelty argument is not in itself a valid
|
||
argument. The fallacy may take two forms: overestimating the new and modern, prematurely and without investigation
|
||
assuming it to be best-case, or underestimating status quo, prematurely and without investigation assuming it to
|
||
be worst-case.
|
||
|
||
Investigation may prove these claims to be true, but it is a fallacy to prematurely conclude this only from the
|
||
general claim that all novelty is good.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Appeal_to_novelty",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias HYPERBOLIC_DISCOUNTING = new()
|
||
{
|
||
Id = new Guid("19a483d0-2c8f-486f-bf9e-619d0df4c916"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Hyperbolic Discounting",
|
||
Description =
|
||
"""
|
||
# Hyperbolic Discounting
|
||
Given two similar rewards, humans show a preference for one that arrives in a more prompt timeframe. Humans are said
|
||
to discount the value of the later reward, by a factor that increases with the length of the delay. In the financial
|
||
world, this process is normally modeled in the form of exponential discounting, a time-consistent model of discounting.
|
||
Many psychological studies have since demonstrated deviations in instinctive preference from the constant discount rate
|
||
assumed in exponential discounting. Hyperbolic discounting is an alternative mathematical model that agrees more closely
|
||
with these findings.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Hyperbolic_discounting",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias RISK_COMPENSATION = new()
|
||
{
|
||
Id = new Guid("10fcc295-02b6-4dbf-b655-f5bcff3c1ca7"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Risk Compensation",
|
||
Description =
|
||
"""
|
||
# Risk Compensation
|
||
Risk compensation is a theory which suggests that people typically adjust their behavior in response to perceived
|
||
levels of risk, becoming more careful where they sense greater risk and less careful if they feel more protected.
|
||
Although usually small in comparison to the fundamental benefits of safety interventions, it may result in a lower
|
||
net benefit than expected or even higher risks.
|
||
|
||
By way of example, it has been observed that motorists drove closer to the vehicle in front when the vehicles were
|
||
fitted with anti-lock brakes. There is also evidence that the risk compensation phenomenon could explain the failure
|
||
of condom distribution programs to reverse HIV prevalence and that condoms may foster disinhibition, with people
|
||
engaging in risky sex both with and without condoms.
|
||
|
||
By contrast, shared space is an urban street design method which consciously aims to increase the level of perceived
|
||
risk and uncertainty, thereby slowing traffic and reducing the number and seriousness of injuries.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Risk_compensation",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias EFFORT_JUSTIFICATION = new()
|
||
{
|
||
Id = new Guid("cff2c74d-a160-4a90-b0b2-10f145b804cb"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Effort Justification",
|
||
Description =
|
||
"""
|
||
# Effort Justification
|
||
Effort justification is an idea and paradigm in social psychology stemming from Leon Festinger's theory of cognitive
|
||
dissonance. Effort justification is a person's tendency to attribute the value of an outcome they put effort into
|
||
achieving as greater than the objective value of the outcome.
|
||
|
||
Cognitive dissonance theory explains changes in people's attitudes or beliefs as the result of an attempt to reduce a
|
||
dissonance (discrepancy) between contradicting ideas or cognitions. In the case of effort justification, there is a
|
||
dissonance between the amount of effort exerted into achieving a goal or completing a task (high effort equalling high
|
||
"cost") and the subjective reward for that effort (lower than was expected for such an effort). By adjusting and increasing
|
||
one's attitude or subjective value of the goal, this dissonance is resolved.
|
||
|
||
One of the first and most classic examples of effort justification is Aronson and Mills's study. A group of young women
|
||
who volunteered to join a discussion group on the topic of the psychology of sex were asked to do a small reading test
|
||
to make sure they were not too embarrassed to talk about sexual-related topics with others. The mild-embarrassment
|
||
condition subjects were asked to read aloud a list of sex-related words such as prostitute or virgin. The
|
||
severe-embarrassment condition subjects were asked to read aloud a list of highly sexual words (e.g. fuck, cock) and
|
||
to read two vivid descriptions of sexual activity taken from contemporary novels. All subjects then listened to a
|
||
recording of a discussion about sexual behavior in animals which was dull and unappealing. When asked to rate the
|
||
group and its members, control and mild-embarrassment groups did not differ, but the severe-embarrassment group's
|
||
ratings were significantly higher. This group, whose initiation process was more difficult (embarrassment equalling
|
||
effort), had to increase their subjective value of the discussion group to resolve the dissonance.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("565616dc-ed84-42af-b9cc-6fa666cc5d66"), // IKEA_EFFECT
|
||
new Guid("ad32d669-fc79-44c9-a570-609e1ccdc799"), // OMISSION_BIAS
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Effort_justification",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias TRAIT_ASCRIPTION_BIAS = new()
|
||
{
|
||
Id = new Guid("4727839d-64c5-4ba4-b044-6b09f14d5a34"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Trait Ascription Bias",
|
||
Description =
|
||
"""
|
||
# Trait Ascription Bias
|
||
Trait ascription bias is the tendency for people to view themselves as relatively variable in terms of personality,
|
||
behavior and mood while viewing others as much more predictable in their personal traits across different situations.
|
||
More specifically, it is a tendency to describe one's own behaviour in terms of situational factors while preferring
|
||
to describe another's behaviour by ascribing fixed dispositions to their personality. This may occur because peoples'
|
||
own internal states are more readily observable and available to them than those of others.
|
||
|
||
This attributional bias intuitively plays a role in the formation and maintenance of stereotypes and prejudice,
|
||
combined with the negativity effect. However, trait ascription and trait-based models of personality remain
|
||
contentious in modern psychology and social science research. Trait ascription bias refers to the situational
|
||
and dispositional evaluation and description of personality traits on a personal level. A similar bias on the
|
||
group level is called the outgroup homogeneity bias.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("ef521fbb-c20b-47c9-87f8-a571a06a03eb"), // NEGATIVITY_BIAS
|
||
new Guid("2cb8514a-c4a2-4cf6-aed7-72d7870ace84"), // BARNUM_EFFECT
|
||
new Guid("f1570784-f8ec-46fd-8bb8-763aef31a04a"), // FUNDAMENTAL_ATTRIBUTION_ERROR
|
||
new Guid("a44a6bcf-b2b8-47f1-84e0-d740af56aa1e"), // ILLUSION_OF_ASYMMETRIC_INSIGHT
|
||
new Guid("f8fd4635-69b3-47be-8243-8c7c6749cae2"), // ILLUSORY_SUPERIORITY
|
||
new Guid("80f9b496-798a-4a1e-a426-815f23b8698e"), // INTROSPECTION_ILLUSION
|
||
new Guid("5ae6f7ec-3be2-47ad-ad75-0ed114f97fe0"), // NAÏVE_CYNICISM
|
||
new Guid("ca2d4f1f-924f-44ae-886b-19240cf2c8c0"), // ULTIMATE_ATTRIBUTION_ERROR
|
||
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Trait_ascription_bias",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias DEFENSIVE_ATTRIBUTION_HYPOTHESIS = new()
|
||
{
|
||
Id = new Guid("5a973490-c19a-43c7-8a01-a26e0d05f275"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Defensive Attribution Hypothesis",
|
||
Description =
|
||
"""
|
||
# Defensive Attribution Hypothesis
|
||
The defensive attribution hypothesis (or bias, theory, or simply defensive attribution) is a social
|
||
psychological term where an observer attributes the causes for a mishap to minimize their fear of
|
||
being a victim or a cause in a similar situation. The attributions of blame are negatively correlated
|
||
to similarities between the observer and the people involved in the mishap, i.e. more responsibility
|
||
is attributed to the people involved who are dissimilar to the observer. Assigning responsibility
|
||
allows the observer to believe that the mishap was controllable and thus preventable.
|
||
|
||
A defensive attribution may also be used to protect the person's self-esteem if, despite everything,
|
||
the mishap does occur, because blame can be assigned to the "other" (person or situation). The use of
|
||
defensive attributions is considered a cognitive bias because an individual will change their beliefs
|
||
about a situation based upon their motivations or desires rather than the factual characteristics of
|
||
the situation.
|
||
|
||
## Sexual assault
|
||
Researchers examining sexual assault have consistently found that male participants blamed rapists less
|
||
than female participants did, and that male participants blamed the rape victims more than female
|
||
participants did. These findings support Shaver's similarity-responsibility hypothesis: male participants,
|
||
who are personally similar to (male) rapists, blame rapists less than female participants who are dissimilar
|
||
to rapists. On the other hand, female participants, who are personally similar to (female) rape victims,
|
||
blame the victims less than male participants.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("efceb4b1-e19f-4997-9f96-1657bb269b2d"), // ATTRIBUTION_BIAS
|
||
new Guid("ad32d669-fc79-44c9-a570-609e1ccdc799"), // OMISSION_BIAS
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Defensive_attribution_hypothesis",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias FUNDAMENTAL_ATTRIBUTION_ERROR = new()
|
||
{
|
||
Id = new Guid("f1570784-f8ec-46fd-8bb8-763aef31a04a"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Fundamental Attribution Error",
|
||
Description =
|
||
"""
|
||
# Fundamental Attribution Error
|
||
In social psychology, the fundamental attribution error (FAE) [a] is a cognitive attribution bias in which
|
||
observers underemphasize situational and environmental factors for the behavior of an actor while overemphasizing
|
||
dispositional or personality factors. In other words, observers tend to overattribute the behaviors of others to
|
||
their personality (e.g., he is late because he's selfish) and underattribute them to the situation or context
|
||
(e.g., he is late because he got stuck in traffic). Although personality traits and predispositions are considered
|
||
to be observable facts in psychology, the fundamental attribution error is an error because it misinterprets their
|
||
effects.
|
||
|
||
The group attribution error (GAE) is identical to the fundamental attribution error, where the bias is shown between
|
||
members of different groups rather than different individuals. The ultimate attribution error is a derivative of the
|
||
FAE and GAE relating to the actions of groups, with an additional layer of self-justification relating to whether
|
||
the action of an individual is representative of the wider group.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("ca2d4f1f-924f-44ae-886b-19240cf2c8c0"), // ULTIMATE_ATTRIBUTION_ERROR
|
||
new Guid("577e79e5-0a53-4c4c-a2ea-d039870bfbb9"), // GROUP_ATTRIBUTION_ERROR
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Fundamental_attribution_error",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias ILLUSION_OF_CONTROL = new()
|
||
{
|
||
Id = new Guid("7fce783e-2120-4aad-9805-2c2a2b937b7d"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Illusion of Control",
|
||
Description =
|
||
"""
|
||
# Illusion of Control
|
||
The illusion of control is the tendency for people to overestimate their ability to control events. It was named
|
||
by U.S. psychologist Ellen Langer and is thought to influence gambling behavior and belief in the paranormal.
|
||
|
||
It is the tendency for people to overestimate their ability to control events, for example, when someone feels a
|
||
sense of control over outcomes that they demonstrably do not influence. The illusion might arise because a person
|
||
lacks direct introspective insight into whether they are in control of events. This has been called the introspection
|
||
illusion. Instead, they may judge their degree of control by a process which is often unreliable. As a result, they see
|
||
themselves as responsible for events to which there is little or no causal link. For example, in one study, college
|
||
students were in a virtual reality setting to treat a fear of heights using an elevator. Those who were told that they
|
||
had control, yet had none, felt as though they had as much control as those who actually did have control over the
|
||
elevator. Those who were led to believe they did not have control said they felt as though they had little control.
|
||
|
||
The illusion is more common in familiar situations, and in situations where the person knows the desired outcome.
|
||
Feedback that emphasizes success rather than failure can increase the effect, while feedback that emphasizes failure
|
||
can decrease or reverse the effect. The illusion is weaker for depressed individuals and is stronger when individuals
|
||
have an emotional need to control the outcome. The illusion is strengthened by stressful and competitive situations,
|
||
including financial trading. Although people are likely to overestimate their control when the situations are heavily
|
||
chance-determined, they also tend to underestimate their control when they actually have it, which runs contrary to
|
||
some theories of the illusion and its adaptiveness. People also showed a higher illusion of control when they were
|
||
allowed to become familiar with a task through practice trials, make their choice before the event happens like
|
||
with throwing dice, and when they can make their choice rather than have it made for them with the same odds.
|
||
People are more likely to show control when they have more answers right at the beginning than at the end,
|
||
even when the people had the same number of correct answers.
|
||
|
||
Being in a position of power enhances the illusion of control, which may lead to overreach in risk taking.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("80f9b496-798a-4a1e-a426-815f23b8698e"), // INTROSPECTION_ILLUSION
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Illusion_of_control",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias ACTOR_OBSERVER_BIAS = new()
|
||
{
|
||
Id = new Guid("5da6dcf4-ed01-4e14-99b0-7a624b16cf17"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Actor-Observer Bias",
|
||
Description =
|
||
"""
|
||
# Actor-Observer Bias
|
||
Actor–observer asymmetry (also actor–observer bias) is a bias one makes when forming attributions about the behavior
|
||
of others or themselves. When people judge their own behavior, they are more likely to attribute their actions to the
|
||
particular situation than to their personality. However, when an observer is explaining the behavior of another person,
|
||
they are more likely to attribute this behavior to the actors' personality rather than to situational factors.
|
||
|
||
Sometimes the actor–observer asymmetry is defined as the fundamental attribution error, which is when people tend to
|
||
explain behavior on the internal, personal characteristics rather than the external factors or situational influences.
|
||
|
||
The specific hypothesis of an actor–observer asymmetry in attribution was originally proposed by Edward Jones and
|
||
Richard Nisbett, where they said that "actors tend to attribute the causes of their behavior to stimuli inherent
|
||
in the situation, while observers tend to attribute behavior to stable dispositions of the actor". Supported by
|
||
initial evidence, the hypothesis was long held as firmly established. However, a meta-analysis of all the published
|
||
tests of the hypothesis between 1971 and 2004 found that there was no actor–observer asymmetry of the sort that had
|
||
been previously proposed. The author of the study interpreted this result not so much as proof that actors and observers
|
||
explained behavior exactly the same way but as evidence that the original hypothesis was fundamentally flawed in the way
|
||
it framed people's explanations of behavior as attributions to either stable dispositions or the situation.
|
||
|
||
Considerations of actor–observer differences can be found in other disciplines as well, such as philosophy (e.g.
|
||
privileged access, incorrigibility), management studies, artificial intelligence, semiotics, anthropology, and
|
||
political science.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("ca2d4f1f-924f-44ae-886b-19240cf2c8c0"), // ULTIMATE_ATTRIBUTION_ERROR
|
||
new Guid("f1570784-f8ec-46fd-8bb8-763aef31a04a"), // FUNDAMENTAL_ATTRIBUTION_ERROR
|
||
new Guid("b57a862b-b490-4d61-96b8-29d548c2eee4"), // POSITIVITY_EFFECT
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Actor%E2%80%93observer_asymmetry",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias SELF_SERVING_BIAS = new()
|
||
{
|
||
Id = new Guid("923ee6c0-2f9c-47fc-a570-339190c1a250"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Self-Serving Bias",
|
||
Description =
|
||
"""
|
||
# Self-Serving Bias
|
||
A self-serving bias is any cognitive or perceptual process that is distorted by the need to maintain and enhance
|
||
self-esteem, or the tendency to perceive oneself in an overly favorable manner. It is the belief that individuals
|
||
tend to ascribe success to their own abilities and efforts, but ascribe failure to external factors. When individuals
|
||
reject the validity of negative feedback, focus on their strengths and achievements but overlook their faults and
|
||
failures, or take more credit for their group's work than they give to other members, they are protecting their
|
||
self-esteem from threat and injury. These cognitive and perceptual tendencies perpetuate illusions and error, but
|
||
they also serve the self's need for esteem. For example, a student who attributes earning a good grade on an exam
|
||
to their own intelligence and preparation but attributes earning a poor grade to the teacher's poor teaching ability
|
||
or unfair test questions might be exhibiting a self-serving bias. Studies have shown that similar attributions are
|
||
made in various situations, such as the workplace, interpersonal relationships, sports, and consumer decisions.
|
||
|
||
Both motivational processes (i.e. self-enhancement, self-presentation) and cognitive processes (i.e. locus of control,
|
||
self-esteem) influence the self-serving bias. There are both cross-cultural (i.e. individualistic and collectivistic
|
||
culture differences) and special clinical population (i.e. depression) considerations within the bias. Much of the
|
||
research on the self-serving bias has used participant self-reports of attribution based on experimental manipulation
|
||
of task outcomes or in naturalistic situations. Some more modern research, however, has shifted focus to physiological
|
||
manipulations, such as emotional inducement and neural activation, in an attempt to better understand the biological
|
||
mechanisms that contribute to the self-serving bias.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("b9c06da1-d2eb-4871-8159-a2a6d25e9eff"), // DUNNING_KRUGER_EFFECT
|
||
new Guid("f1570784-f8ec-46fd-8bb8-763aef31a04a"), // FUNDAMENTAL_ATTRIBUTION_ERROR
|
||
new Guid("f8fd4635-69b3-47be-8243-8c7c6749cae2"), // ILLUSORY_SUPERIORITY
|
||
new Guid("ad32d669-fc79-44c9-a570-609e1ccdc799"), // OMISSION_BIAS
|
||
new Guid("7bf44f8f-a4b0-404c-8f15-8ca6e3322d32"), // OPTIMISM_BIAS
|
||
new Guid("b821d449-64e5-4c0a-9d5a-3fda609a9b86"), // OVERCONFIDENCE_EFFECT
|
||
new Guid("e36f82b7-43dd-4073-99d9-c33073007185"), // MORAL_CREDENTIAL_EFFECT
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Self-serving_bias",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias OPTIMISM_BIAS = new()
|
||
{
|
||
Id = new Guid("7bf44f8f-a4b0-404c-8f15-8ca6e3322d32"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Optimism Bias",
|
||
Description =
|
||
"""
|
||
# Optimism Bias
|
||
Optimism bias or optimistic bias is a cognitive bias that causes someone to believe that they themselves
|
||
are less likely to experience a negative event. It is also known as unrealistic optimism or comparative optimism.
|
||
|
||
Optimism bias is common and transcends gender, ethnicity, nationality, and age. However, autistic people are less
|
||
susceptible to this kind of biases. Optimistic biases have also reported in other animals, such as rats and birds.
|
||
|
||
Four factors can cause a person to be optimistically biased: their desired end state, their cognitive mechanisms,
|
||
the information they have about themselves versus others, and overall mood. The optimistic bias is seen in a number
|
||
of situations. For example: people believing that they are less at risk of being a crime victim, smokers believing
|
||
that they are less likely to contract lung cancer or disease than other smokers, first-time bungee jumpers believing
|
||
that they are less at risk of an injury than other jumpers, or traders who think they are less exposed to potential
|
||
losses in the markets.
|
||
|
||
Although the optimism bias occurs for both positive events (such as believing oneself to be more financially successful
|
||
than others) and negative events (such as being less likely to have a drinking problem), there is more research and
|
||
evidence suggesting that the bias is stronger for negative events (the valence effect). Different consequences result
|
||
from these two types of events: positive events often lead to feelings of well being and self-esteem, while negative
|
||
events lead to consequences involving more risk, such as engaging in risky behaviors and not taking precautionary
|
||
measures for safety.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("67041978-ac8e-4254-ae2c-509e7301619f"), // PESSIMISM_BIAS
|
||
new Guid("7fce783e-2120-4aad-9805-2c2a2b937b7d"), // ILLUSION_OF_CONTROL
|
||
new Guid("f8fd4635-69b3-47be-8243-8c7c6749cae2"), // ILLUSORY_SUPERIORITY
|
||
new Guid("ef521fbb-c20b-47c9-87f8-a571a06a03eb"), // NEGATIVITY_BIAS
|
||
new Guid("b57a862b-b490-4d61-96b8-29d548c2eee4"), // POSITIVITY_EFFECT
|
||
new Guid("923ee6c0-2f9c-47fc-a570-339190c1a250"), // SELF_SERVING_BIAS
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Optimism_bias",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias EGOCENTRIC_BIAS = new()
|
||
{
|
||
Id = new Guid("953746dc-ce10-4e3b-8f9e-9246de63f531"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Egocentric Bias",
|
||
Description =
|
||
"""
|
||
# Egocentric Bias
|
||
Egocentric bias is the tendency to rely too heavily on one's own perspective and/or have a higher opinion of
|
||
oneself than reality. It appears to be the result of the psychological need to satisfy one's ego and to be
|
||
advantageous for memory consolidation. Research has shown that experiences, ideas, and beliefs are more easily
|
||
recalled when they match one's own, causing an egocentric outlook. Michael Ross and Fiore Sicoly first identified
|
||
this cognitive bias in their 1979 paper, "Egocentric Biases in Availability and Attribution". Egocentric bias is
|
||
referred to by most psychologists as a general umbrella term under which other related phenomena fall.
|
||
|
||
The effects of egocentric bias can differ based on personal characteristics, such as age and the number of
|
||
languages one speaks. Thus far, there have been many studies focusing on specific implications of egocentric
|
||
bias in different contexts. Research on collaborative group tasks have emphasized that people view their own
|
||
contributions differently than they view that of others. Other areas of research have been aimed at studying
|
||
how mental health patients display egocentric bias, and at the relationship between egocentric bias and voter
|
||
distribution. These types of studies surrounding egocentric bias usually involve written or verbal questionnaires,
|
||
based on the subject's personal life or their decision in various hypothetical scenarios.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("923ee6c0-2f9c-47fc-a570-339190c1a250"), // SELF_SERVING_BIAS
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Egocentric_bias",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias DUNNING_KRUGER_EFFECT = new()
|
||
{
|
||
Id = new Guid("b9c06da1-d2eb-4871-8159-a2a6d25e9eff"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Dunning-Kruger Effect",
|
||
Description =
|
||
"""
|
||
# Dunning-Kruger Effect
|
||
The Dunning–Kruger effect is a cognitive bias in which people with limited competence in a particular domain overestimate
|
||
their abilities. It was first described by Justin Kruger and David Dunning in 1999. Some researchers also include the
|
||
opposite effect for high performers: their tendency to underestimate their skills. In popular culture, the Dunning–Kruger
|
||
effect is often misunderstood as a claim about general overconfidence of people with low intelligence instead of specific
|
||
overconfidence of people unskilled at a particular task.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("f8fd4635-69b3-47be-8243-8c7c6749cae2"), // ILLUSORY_SUPERIORITY
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Dunning%E2%80%93Kruger_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias HARD_EASY_EFFECT = new()
|
||
{
|
||
Id = new Guid("07f0c252-1d97-4207-8000-8e4d8800fb42"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Hard-Easy Effect",
|
||
Description =
|
||
"""
|
||
# Hard-Easy Effect
|
||
The hard–easy effect is a cognitive bias that manifests itself as a tendency to overestimate the probability of
|
||
one's success at a task perceived as hard, and to underestimate the likelihood of one's success at a task perceived
|
||
as easy. The hard-easy effect takes place, for example, when individuals exhibit a degree of underconfidence in
|
||
answering relatively easy questions and a degree of overconfidence in answering relatively difficult questions.
|
||
"Hard tasks tend to produce overconfidence but worse-than-average perceptions," reported Katherine A. Burson,
|
||
Richard P. Larrick, and Jack B. Soll in a 2005 study, "whereas easy tasks tend to produce underconfidence and
|
||
better-than-average effects."
|
||
|
||
The hard-easy effect falls under the umbrella of "social comparison theory", which was originally formulated by
|
||
Leon Festinger in 1954. Festinger argued that individuals are driven to evaluate their own opinions and abilities
|
||
accurately, and social comparison theory explains how individuals carry out those evaluations by comparing themselves
|
||
to others.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("b821d449-64e5-4c0a-9d5a-3fda609a9b86"), // OVERCONFIDENCE_EFFECT
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Hard%E2%80%93easy_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias FALSE_CONSENSUS_EFFECT = new()
|
||
{
|
||
Id = new Guid("bc0dc6d3-5115-4def-91ae-a38aebed185e"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "False Consensus Effect",
|
||
Description =
|
||
"""
|
||
# False Consensus Effect
|
||
In psychology, the false consensus effect, also known as consensus bias, is a pervasive cognitive bias that causes
|
||
people to "see their own behavioral choices and judgments as relatively common and appropriate to existing
|
||
circumstances". In other words, they assume that their personal qualities, characteristics, beliefs, and actions
|
||
are relatively widespread through the general population.
|
||
|
||
This false consensus is significant because it increases self-esteem (overconfidence effect). It can be derived from
|
||
a desire to conform and be liked by others in a social environment. This bias is especially prevalent in group
|
||
settings where one thinks the collective opinion of their own group matches that of the larger population. Since
|
||
the members of a group reach a consensus and rarely encounter those who dispute it, they tend to believe that
|
||
everybody thinks the same way. The false-consensus effect is not restricted to cases where people believe that
|
||
their values are shared by the majority, but it still manifests as an overestimate of the extent of their belief.
|
||
|
||
Additionally, when confronted with evidence that a consensus does not exist, people often assume that those who
|
||
do not agree with them are defective in some way. There is no single cause for this cognitive bias; the
|
||
availability heuristic, self-serving bias, and naïve realism have been suggested as at least partial underlying
|
||
factors. The bias may also result, at least in part, from non-social stimulus-reward associations. Maintenance
|
||
of this cognitive bias may be related to the tendency to make decisions with relatively little information.
|
||
When faced with uncertainty and a limited sample from which to make decisions, people often "project"
|
||
themselves onto the situation. When this personal knowledge is used as input to make generalizations,
|
||
it often results in the false sense of being part of the majority.
|
||
""",
|
||
|
||
Related = [
|
||
new Guid("b821d449-64e5-4c0a-9d5a-3fda609a9b86"), // OVERCONFIDENCE_EFFECT
|
||
new Guid("d749ce96-32f3-4c3d-86f7-26ff4edabe4a"), // AVAILABILITY_HEURISTIC
|
||
new Guid("923ee6c0-2f9c-47fc-a570-339190c1a250"), // SELF_SERVING_BIAS
|
||
new Guid("f0ad095e-8e9c-4bfb-855e-11fb5dd58cea"), // NAÏVE_REALISM
|
||
],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/False_consensus_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias THIRD_PERSON_EFFECT = new()
|
||
{
|
||
Id = new Guid("b9186d75-3362-4dd4-a3ec-4345a04161c9"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Third-Person Effect",
|
||
Description =
|
||
"""
|
||
# Third-Person Effect
|
||
The third-person effect hypothesis predicts that people tend to perceive that mass media messages have a greater
|
||
effect on others than on themselves, based on personal biases. The third-person effect manifests itself through
|
||
an individual's overestimation of the effect of a mass communicated message on the generalized other, or an
|
||
underestimation of the effect of a mass communicated message on themselves.
|
||
|
||
These types of perceptions stem from a self-motivated social desirability (not feeling influenced by mass messages
|
||
promotes self-esteem), a social-distance corollary (choosing to dissociate oneself from the others who may be
|
||
influenced), and a perceived exposure to a message (others choose to be influenced by persuasive communication).
|
||
Other names for the effect are "Third-person perception" and "Web Third-person effect". From 2015, the effect is
|
||
named "Web Third-person effect" when it is verified in social media, media websites, blogs and in websites in general.
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Third-person_effect",
|
||
],
|
||
};
|
||
|
||
private static readonly Bias SOCIAL_DESIRABILITY_BIAS = new()
|
||
{
|
||
Id = new Guid("a378b725-8bf9-4213-a875-326426d5c759"),
|
||
Category = BiasCategory.NEED_TO_ACT_FAST,
|
||
Name = "Social-Desirability Bias",
|
||
Description =
|
||
"""
|
||
# Social-Desirability Bias
|
||
In social science research, social-desirability bias is a type of response bias that is the tendency of survey
|
||
respondents to answer questions in a manner that will be viewed favorably by others. It can take the form of
|
||
over-reporting "good behavior" or under-reporting "bad", or undesirable behavior. The tendency poses a serious
|
||
problem with conducting research with self-reports. This bias interferes with the interpretation of average
|
||
tendencies as well as individual differences.
|
||
|
||
Topics where socially desirable responding (SDR) is of special concern are self-reports of abilities, personality,
|
||
sexual behavior, and drug use. When confronted with the question "How often do you masturbate?," for example,
|
||
respondents may be pressured by a social taboo against masturbation, and either under-report the frequency or
|
||
avoid answering the question. Therefore, the mean rates of masturbation derived from self-report surveys are
|
||
likely to be severely underestimated.
|
||
|
||
When confronted with the question, "Do you use drugs/illicit substances?" the respondent may be influenced by
|
||
the fact that controlled substances, including the more commonly used marijuana, are generally illegal.
|
||
Respondents may feel pressured to deny any drug use or rationalize it, e.g. "I only smoke marijuana when my
|
||
friends are around." The bias can also influence reports of number of sexual partners. In fact, the bias
|
||
may operate in opposite directions for different subgroups: Whereas men tend to inflate the numbers, women
|
||
tend to underestimate theirs. In either case, the mean reports from both groups are likely to be distorted
|
||
by social desirability bias.
|
||
|
||
Other topics that are sensitive to social-desirability bias include:
|
||
|
||
- Self-reported personality traits will correlate strongly with social desirability bias
|
||
- Personal income and earnings, often inflated when low and deflated when high
|
||
- Feelings of low self-worth and/or powerlessness, often denied
|
||
- Excretory functions, often approached uncomfortably, if discussed at all
|
||
- Compliance with medicinal-dosing schedules, often inflated
|
||
- Family planning, including use of contraceptives and abortion
|
||
- Religion, often either avoided or uncomfortably approached
|
||
- Patriotism, either inflated or, if denied, done so with a fear of other party's judgment
|
||
- Bigotry and intolerance, often denied, even if it exists within the responder
|
||
- Intellectual achievements, often inflated
|
||
- Physical appearance, either inflated or deflated
|
||
- Acts of real or imagined physical violence, often denied
|
||
- Indicators of charity or "benevolence," often inflated
|
||
- Illegal acts, often denied
|
||
- Voter turnout
|
||
""",
|
||
|
||
Related = [],
|
||
Links =
|
||
[
|
||
"https://en.wikipedia.org/wiki/Social-desirability_bias",
|
||
],
|
||
};
|
||
|
||
#endregion
|
||
|
||
public static readonly IReadOnlyDictionary<Guid, Bias> ALL_BIAS = new Dictionary<Guid, Bias>
|
||
{
|
||
{ SOCIAL_DESIRABILITY_BIAS.Id, SOCIAL_DESIRABILITY_BIAS },
|
||
{ THIRD_PERSON_EFFECT.Id, THIRD_PERSON_EFFECT },
|
||
{ FALSE_CONSENSUS_EFFECT.Id, FALSE_CONSENSUS_EFFECT },
|
||
{ HARD_EASY_EFFECT.Id, HARD_EASY_EFFECT },
|
||
{ DUNNING_KRUGER_EFFECT.Id, DUNNING_KRUGER_EFFECT },
|
||
{ EGOCENTRIC_BIAS.Id, EGOCENTRIC_BIAS },
|
||
{ OPTIMISM_BIAS.Id, OPTIMISM_BIAS },
|
||
{ SELF_SERVING_BIAS.Id, SELF_SERVING_BIAS },
|
||
{ ACTOR_OBSERVER_BIAS.Id, ACTOR_OBSERVER_BIAS },
|
||
{ ILLUSION_OF_CONTROL.Id, ILLUSION_OF_CONTROL },
|
||
{ FUNDAMENTAL_ATTRIBUTION_ERROR.Id, FUNDAMENTAL_ATTRIBUTION_ERROR },
|
||
{ DEFENSIVE_ATTRIBUTION_HYPOTHESIS.Id, DEFENSIVE_ATTRIBUTION_HYPOTHESIS },
|
||
{ TRAIT_ASCRIPTION_BIAS.Id, TRAIT_ASCRIPTION_BIAS },
|
||
{ EFFORT_JUSTIFICATION.Id, EFFORT_JUSTIFICATION },
|
||
{ RISK_COMPENSATION.Id, RISK_COMPENSATION },
|
||
{ HYPERBOLIC_DISCOUNTING.Id, HYPERBOLIC_DISCOUNTING },
|
||
{ APPEAL_TO_NOVELTY.Id, APPEAL_TO_NOVELTY },
|
||
{ IDENTIFIABLE_VICTIM_EFFECT.Id, IDENTIFIABLE_VICTIM_EFFECT },
|
||
{ SUNK_COST_FALLACY.Id, SUNK_COST_FALLACY },
|
||
{ ESCALATION_OF_COMMITMENT.Id, ESCALATION_OF_COMMITMENT },
|
||
{ GENERATION_EFFECT.Id, GENERATION_EFFECT },
|
||
{ LOSS_AVERSION.Id, LOSS_AVERSION },
|
||
{ IKEA_EFFECT.Id, IKEA_EFFECT },
|
||
{ UNIT_BIAS.Id, UNIT_BIAS },
|
||
{ ZERO_RISK_BIAS.Id, ZERO_RISK_BIAS },
|
||
{ DISPOSITION_EFFECT.Id, DISPOSITION_EFFECT },
|
||
{ CERTAINTY_EFFECT.Id, CERTAINTY_EFFECT },
|
||
{ PSEUDOCERTAINTY_EFFECT.Id, PSEUDOCERTAINTY_EFFECT },
|
||
{ PROCESSING_DIFFICULTY_EFFECT.Id, PROCESSING_DIFFICULTY_EFFECT },
|
||
{ ENDOWMENT_EFFECT.Id, ENDOWMENT_EFFECT },
|
||
{ BELIEF_PERSEVERANCE.Id, BELIEF_PERSEVERANCE },
|
||
{ SYSTEM_JUSTIFICATION.Id, SYSTEM_JUSTIFICATION },
|
||
{ REVERSE_PSYCHOLOGY.Id, REVERSE_PSYCHOLOGY },
|
||
{ REACTANCE.Id, REACTANCE },
|
||
{ DECOY_EFFECT.Id, DECOY_EFFECT },
|
||
{ SOCIAL_COMPARISON_BIAS.Id, SOCIAL_COMPARISON_BIAS },
|
||
{ STATUS_QUO_BIAS.Id, STATUS_QUO_BIAS },
|
||
{ AMBIGUITY_EFFECT.Id, AMBIGUITY_EFFECT },
|
||
{ INFORMATION_BIAS.Id, INFORMATION_BIAS },
|
||
{ BELIEF_BIAS.Id, BELIEF_BIAS },
|
||
{ RHYME_AS_REASON_EFFECT.Id, RHYME_AS_REASON_EFFECT },
|
||
{ PARKINSONS_LAW_OF_TRIVIALITY.Id, PARKINSONS_LAW_OF_TRIVIALITY },
|
||
{ DELMORE_EFFECT.Id, DELMORE_EFFECT },
|
||
{ CONJUNCTION_FALLACY.Id, CONJUNCTION_FALLACY },
|
||
{ OCCAMS_RAZOR.Id, OCCAMS_RAZOR },
|
||
{ LESS_IS_BETTER_EFFECT.Id, LESS_IS_BETTER_EFFECT},
|
||
{ HINDSIGHT_BIAS.Id, HINDSIGHT_BIAS },
|
||
{ OUTCOME_BIAS.Id, OUTCOME_BIAS },
|
||
{ MORAL_LUCK.Id, MORAL_LUCK },
|
||
{ DECLINISM.Id, DECLINISM },
|
||
{ PESSIMISM_BIAS.Id, PESSIMISM_BIAS },
|
||
{ PLANNING_FALLACY.Id, PLANNING_FALLACY },
|
||
{ TIME_SAVING_BIAS.Id, TIME_SAVING_BIAS },
|
||
{ PRO_INNOVATION_BIAS.Id, PRO_INNOVATION_BIAS },
|
||
{ IMPACT_BIAS.Id, IMPACT_BIAS },
|
||
{ PROJECTION_BIAS.Id, PROJECTION_BIAS },
|
||
{ ROSY_RETROSPECTION.Id, ROSY_RETROSPECTION },
|
||
{ TELESCOPING_EFFECT.Id, TELESCOPING_EFFECT },
|
||
{ ILLUSION_OF_ASYMMETRIC_INSIGHT.Id, ILLUSION_OF_ASYMMETRIC_INSIGHT },
|
||
{ ILLUSION_OF_EXTERNAL_AGENCY.Id, ILLUSION_OF_EXTERNAL_AGENCY },
|
||
{ EXTRINSIC_INCENTIVE_BIAS.Id, EXTRINSIC_INCENTIVE_BIAS },
|
||
{ SPOTLIGHT_EFFECT.Id, SPOTLIGHT_EFFECT },
|
||
{ CURSE_OF_KNOWLEDGE.Id, CURSE_OF_KNOWLEDGE },
|
||
{ ILLUSION_OF_TRANSPARENCY.Id, ILLUSION_OF_TRANSPARENCY },
|
||
{ MILLERS_LAW.Id, MILLERS_LAW },
|
||
{ DENOMINATION_EFFECT.Id, DENOMINATION_EFFECT },
|
||
{ SUBADDITIVITY_EFFECT.Id, SUBADDITIVITY_EFFECT },
|
||
{ SURVIVORSHIP_BIAS.Id, SURVIVORSHIP_BIAS },
|
||
{ ZERO_SUM_BIAS.Id, ZERO_SUM_BIAS },
|
||
{ NORMALCY_BIAS.Id, NORMALCY_BIAS },
|
||
{ APPEAL_TO_POSSIBILITY.Id, APPEAL_TO_POSSIBILITY },
|
||
{ MENTAL_ACCOUNTING.Id, MENTAL_ACCOUNTING },
|
||
{ WELL_TRAVELLED_ROAD_EFFECT.Id, WELL_TRAVELLED_ROAD_EFFECT },
|
||
{ REACTIVE_DEVALUATION.Id, REACTIVE_DEVALUATION },
|
||
{ NOT_INVENTED_HERE.Id, NOT_INVENTED_HERE },
|
||
{ POSITIVITY_EFFECT.Id, POSITIVITY_EFFECT },
|
||
{ CHEERLEADER_EFFECT.Id, CHEERLEADER_EFFECT },
|
||
{ CROSS_RACE_EFFECT.Id, CROSS_RACE_EFFECT },
|
||
{ OUT_GROUP_HOMOGENEITY.Id, OUT_GROUP_HOMOGENEITY },
|
||
{ PLACEBO_EFFECT.Id, PLACEBO_EFFECT },
|
||
{ BANDWAGON_EFFECT.Id, BANDWAGON_EFFECT },
|
||
{ AUTOMATION_BIAS.Id, AUTOMATION_BIAS },
|
||
{ AUTHORITY_BIAS.Id, AUTHORITY_BIAS },
|
||
{ ARGUMENT_FROM_FALLACY.Id, ARGUMENT_FROM_FALLACY },
|
||
{ JUST_WORLD_FALLACY.Id, JUST_WORLD_FALLACY },
|
||
{ MORAL_CREDENTIAL_EFFECT.Id, MORAL_CREDENTIAL_EFFECT },
|
||
{ FUNCTIONAL_FIXEDNESS.Id, FUNCTIONAL_FIXEDNESS },
|
||
{ ESSENTIALISM.Id, ESSENTIALISM },
|
||
{ STEREOTYPING.Id, STEREOTYPING },
|
||
{ IN_GROUP_FAVORITISM.Id, IN_GROUP_FAVORITISM },
|
||
{ ULTIMATE_ATTRIBUTION_ERROR.Id, ULTIMATE_ATTRIBUTION_ERROR },
|
||
{ HOSTILE_ATTRIBUTION_BIAS.Id, HOSTILE_ATTRIBUTION_BIAS },
|
||
{ ATTRIBUTION_BIAS.Id, ATTRIBUTION_BIAS },
|
||
{ GROUP_ATTRIBUTION_ERROR.Id, GROUP_ATTRIBUTION_ERROR },
|
||
{ ANTHROPOMORPHISM.Id, ANTHROPOMORPHISM },
|
||
{ APOPHENIA.Id, APOPHENIA },
|
||
{ PAREIDOLIA.Id, PAREIDOLIA },
|
||
{ ILLUSORY_CORRELATION.Id, ILLUSORY_CORRELATION },
|
||
{ HOT_HAND_FALLACY.Id, HOT_HAND_FALLACY },
|
||
{ GAMBLERS_FALLACY.Id, GAMBLERS_FALLACY },
|
||
{ RECENCY_ILLUSION.Id, RECENCY_ILLUSION },
|
||
{ MASKED_MAN_FALLACY.Id, MASKED_MAN_FALLACY },
|
||
{ WYSIATI.Id, WYSIATI },
|
||
{ ILLUSION_OF_VALIDITY.Id, ILLUSION_OF_VALIDITY },
|
||
{ ANECDOTAL_FALLACY.Id, ANECDOTAL_FALLACY },
|
||
{ NEGLECT_OF_PROBABILITY.Id, NEGLECT_OF_PROBABILITY },
|
||
{ INSENSITIVITY_TO_SAMPLE_SIZE.Id, INSENSITIVITY_TO_SAMPLE_SIZE },
|
||
{ CLUSTERING_ILLUSION.Id, CLUSTERING_ILLUSION },
|
||
{ CONFABULATION.Id, CONFABULATION },
|
||
{ NAÏVE_REALISM.Id, NAÏVE_REALISM },
|
||
{ NAÏVE_CYNICISM.Id, NAÏVE_CYNICISM },
|
||
{ OVERCONFIDENCE_EFFECT.Id, OVERCONFIDENCE_EFFECT },
|
||
{ ILLUSORY_SUPERIORITY.Id, ILLUSORY_SUPERIORITY },
|
||
{ INTROSPECTION_ILLUSION.Id, INTROSPECTION_ILLUSION },
|
||
{ BIAS_BLIND_SPOT.Id, BIAS_BLIND_SPOT },
|
||
{ SEMMELWEIS_REFLEX.Id, SEMMELWEIS_REFLEX },
|
||
{ CONTINUED_INFLUENCE_EFFECT.Id, CONTINUED_INFLUENCE_EFFECT },
|
||
{ BARNUM_EFFECT.Id, BARNUM_EFFECT },
|
||
{ SUBJECTIVE_VALIDATION.Id, SUBJECTIVE_VALIDATION },
|
||
{ OSTRICH_EFFECT.Id, OSTRICH_EFFECT },
|
||
{ OBSERVER_EXPECTANCY_EFFECT.Id, OBSERVER_EXPECTANCY_EFFECT },
|
||
{ SELECTIVE_PERCEPTION.Id, SELECTIVE_PERCEPTION },
|
||
{ CHOICE_SUPPORTIVE_BIAS.Id, CHOICE_SUPPORTIVE_BIAS },
|
||
{ CONGRUENCE_BIAS.Id, CONGRUENCE_BIAS },
|
||
{ CONFIRMATION_BIAS.Id, CONFIRMATION_BIAS },
|
||
{ WEBER_FECHNER_LAW.Id, WEBER_FECHNER_LAW },
|
||
{ MONEY_ILLUSION.Id, MONEY_ILLUSION },
|
||
{ FRAMING_EFFECT.Id, FRAMING_EFFECT },
|
||
{ FOCUSING_EFFECT.Id, FOCUSING_EFFECT },
|
||
{ DISTINCTION_BIAS.Id, DISTINCTION_BIAS },
|
||
{ CONTRAST_EFFECT.Id, CONTRAST_EFFECT },
|
||
{ CONSERVATISM_BIAS.Id, CONSERVATISM_BIAS },
|
||
{ ANCHORING_EFFECT.Id, ANCHORING_EFFECT },
|
||
{ SELF_REFERENCE_EFFECT.Id, SELF_REFERENCE_EFFECT },
|
||
{ PICTURE_SUPERIORITY_EFFECT.Id, PICTURE_SUPERIORITY_EFFECT },
|
||
{ VON_RESTORFF_EFFECT.Id, VON_RESTORFF_EFFECT },
|
||
{ HUMOUR_EFFECT.Id, HUMOUR_EFFECT },
|
||
{ BIZARRENESS_EFFECT.Id, BIZARRENESS_EFFECT },
|
||
{ BASE_RATE_FALLACY.Id, BASE_RATE_FALLACY },
|
||
{ OMISSION_BIAS.Id, OMISSION_BIAS},
|
||
{ HOT_COLD_EMPATHY_GAP.Id, HOT_COLD_EMPATHY_GAP },
|
||
{ FREQUENCY_ILLUSION.Id, FREQUENCY_ILLUSION },
|
||
{ CONTEXT_DEPENDENT_MEMORY.Id, CONTEXT_DEPENDENT_MEMORY },
|
||
{ STATE_DEPENDENT_MEMORY.Id, STATE_DEPENDENT_MEMORY },
|
||
{ CUE_DEPENDENT_FORGETTING.Id, CUE_DEPENDENT_FORGETTING },
|
||
{ CONTEXT_EFFECT.Id, CONTEXT_EFFECT },
|
||
{ MERE_EXPOSURE_EFFECT.Id, MERE_EXPOSURE_EFFECT },
|
||
{ ILLUSORY_TRUTH_EFFECT.Id, ILLUSORY_TRUTH_EFFECT },
|
||
{ ATTENTIONAL_BIAS.Id, ATTENTIONAL_BIAS },
|
||
{ AVAILABILITY_HEURISTIC.Id, AVAILABILITY_HEURISTIC },
|
||
{ MODALITY_EFFECT.Id, MODALITY_EFFECT },
|
||
{ MEMORY_INHIBITION.Id, MEMORY_INHIBITION },
|
||
{ PRIMACY_EFFECT.Id, PRIMACY_EFFECT },
|
||
{ RECENCY_EFFECT.Id, RECENCY_EFFECT },
|
||
{ PART_LIST_CUING.Id, PART_LIST_CUING },
|
||
{ SERIAL_POSITION_EFFECT.Id, SERIAL_POSITION_EFFECT },
|
||
{ SUFFIX_EFFECT.Id, SUFFIX_EFFECT },
|
||
{ LEVELS_OF_PROCESSING_EFFECT.Id, LEVELS_OF_PROCESSING_EFFECT },
|
||
{ ABSENT_MINDEDNESS.Id, ABSENT_MINDEDNESS },
|
||
{ TESTING_EFFECT.Id, TESTING_EFFECT },
|
||
{ NEXT_IN_LINE_EFFECT.Id, NEXT_IN_LINE_EFFECT },
|
||
{ GOOGLE_EFFECT.Id, GOOGLE_EFFECT },
|
||
{ TIP_OF_THE_TONGUE_PHENOMENON.Id, TIP_OF_THE_TONGUE_PHENOMENON },
|
||
{ SUGGESTIBILITY.Id, SUGGESTIBILITY },
|
||
{ SPACING_EFFECT.Id, SPACING_EFFECT },
|
||
{ MISATTRIBUTION_OF_MEMORY.Id, MISATTRIBUTION_OF_MEMORY },
|
||
{ LIST_LENGTH_EFFECT.Id, LIST_LENGTH_EFFECT },
|
||
{ MISINFORMATION_EFFECT.Id, MISINFORMATION_EFFECT },
|
||
{ LEVELING_AND_SHARPENING.Id, LEVELING_AND_SHARPENING },
|
||
{ PEAK_END_RULE.Id, PEAK_END_RULE },
|
||
{ FADING_AFFECT_BIAS.Id, FADING_AFFECT_BIAS },
|
||
{ NEGATIVITY_BIAS.Id, NEGATIVITY_BIAS },
|
||
{ PREJUDICE.Id, PREJUDICE },
|
||
{ IMPLICIT_STEREOTYPES.Id, IMPLICIT_STEREOTYPES },
|
||
{ IMPLICIT_ASSOCIATIONS.Id, IMPLICIT_ASSOCIATIONS },
|
||
};
|
||
|
||
public static Bias GetRandomBias(IList<int> usedBias)
|
||
{
|
||
if(usedBias.Count >= ALL_BIAS.Count)
|
||
usedBias.Clear();
|
||
|
||
var randomBiasIndex = Random.Shared.Next(0, ALL_BIAS.Count);
|
||
while(usedBias.Contains(randomBiasIndex))
|
||
randomBiasIndex = Random.Shared.Next(0, ALL_BIAS.Count);
|
||
|
||
usedBias.Add(randomBiasIndex);
|
||
return ALL_BIAS.Values.ElementAt(randomBiasIndex);
|
||
}
|
||
} |