A Deep Dive into Using Algorithms to Detect Preferences from Brain Waves
23rd June 2021 • Human Factors Cast Deep Dives • Human Factors Cast
00:00:00 00:13:49

Share Episode

Transcripts

It’s Wednesday again, and we know what that means: time for another deep dive! This week we will be diving into how we develop preferences, algorithmic bias, and a brief dip into what all of this might mean for our concept of free will.

How Do Our Preferences Form?

It is generally understood that different people have different preferences: one person might prefer chocolate ice cream, another might prefer vanilla. One person might prefer the cold, another might prefer the heat. The same person might prefer coffee in the morning and tea in the afternoon. What makes us have different preferences?

i was a mathematician who, in:

A literature review in:

r preferences is a study from:

Part of that context includes social factors. Economists have developed terminology for two types of preferences that come from external pressures: conspicuous consumption and information cascade. Conspicuous consumption is when we make choices (usually expensive ones) in order to build the image that we want to present to the world (e.g. We want to project that we are eco-friendly, so we drive an electric car, carry a reusable water bottle, and wear clothing made from sustainable fabrics). An information cascade is when so many other people are doing something, we go along with it (e.g. We may not need or particularly want a fidget spinner, but everyone is talking about them and buying them, so we buy one).

Racial Bias in Algorithms

Although we may not know exactly how they work, most of us are aware that much of our online experience is guided through algorithms that predict what we will be interested in. Social media algorithms suggest new videos to watch or people to follow, advertising algorithms suggest products to buy, search engines use a collection of algorithms to suggest results that will appeal to you - algorithms are in use everywhere.

Unfortunately, one aspect of algorithms that is still being figured out is how to get rid of racial bias. While algorithms are typically not purposefully designed to have a racial bias, oftentimes what seems like a small bit of programming results in larger racially discriminatory consequences.

A:

- An algorithm developed by the American Heart Association to determine a heart failure risk score has led to more non-Black patients being referred to specialized care even when symptoms and medical history are the same

- An algorithm for a risk calculator used by thoracic surgeons predicts a higher risk of death and other complications from chest surgery for Black patients, leading to physicians being less likely to recommend them for life-saving procedures

- An algorithm that measures creatine as an indicator of kidney function adjusts creatine measurement levels in Black patients lower, potentially preventing them from receiving proper care

- An algorithm that is used for kidney transplants says that kidneys from Black kidney donors are more likely to fail, which ultimately leads to a reduced number of kidneys available to Black patients who need kidney transplants

- An algorithm that is used to determine if a mother needs to have a cesarean delivery because vaginal birth would be too dangerous applies a higher risk to women of color, setting them up for multiple cesarean deliveries (which is considered increasingly dangerous)

- An algorithm used by an online breast cancer screening tool calculates a lower risk for getting breast cancer for women of color even when all other risk factors are the same

- Algorithmic bias is a huge problem in healthcare, and it is important to keep in mind that it extends beyond the healthcare setting, too. Racial bias in social media algorithms results in Black adolescents facing, on average, over five racial discrimination experiences per day. A computer that learned English by scraping the internet became racist and sexist.

An often-cited:

In an article about racial bias in algorithms used for risk assessment in the judicial system, ProPublica explored how algorithms in use in the criminal justice system discriminate against Black defendants. In an analysis they performed on data from Broward County, Florida, ProPublica found that the COMPAS algorithm wrongly labeled Black defendants as future criminals at almost twice the rate of White defendants, and White defendants were mislabeled as low risk more often than Black defendants.

There is clearly a major, ongoing issue with racial bias in algorithms across all fields. This will need to be noted and addressed in any new algorithms to ensure they can benefit all potential end-users and not perpetuate harmful practices.

If Our Preferences Can Be Predicted, Do We Have Free Will?

Free will and whether or not we have it is a subject that truly deserves a deep dive all its own. While most of us have a general idea of what we think free will is - usually along the lines of our unique ability to be in control of our own actions, thoughts, and destiny - philosophers tend to disagree on a specific definition and method of measurement. There is also significant debate about if free will is simply an illusion.

Some in the psychological community take the position that our behaviors and thoughts are merely the collateral output of the physical happenings in the brain, the byproduct of neurons firing in reaction to stimuli, and mental processes have no influence over physical events. This is called epiphenomenalism.

Benjamin Libet's famous experiments seemed to prove epiphenomenalism, that we do not actually have free will because he found that our brain sent signals to start a behavior before we were aware of having decided to do the behavior. However, there is significant criticism of his findings. Al Mele and others have pointed out that Libet’s experiments did not typically look at events that were salient or important to an individual, the findings do not translate well to the real world, replicability of these findings is low or controversial, actions can be preceded by many things (e.g. urges, wants, intentions), we may plan an action for the future, and our intentions may be specific or relatively unspecific.

It seems that there is not enough data to say that we always have free will, but also not enough to say that we never have it.

What This Means for an Algorithm That Can Determine Our Preferences

A machine that can read our brain waves to determine what our preferences are could have many benefits, as we talked about on the show. It also may be seen as providing more support for the concept of epiphenomenalism: maybe our preferences are not the result of conscious thought, but are just the natural result of how our brains physically work.

If our preferences are so easy to determine by reading brain activity, are they really under our control? Do we really choose them, or is our body programmed to prefer one thing over another?

The idea that we do not have free will, that our actions are predetermined, can lead people to feel angry, decrease their desire to help others, and diminish any feelings of gratitude they may have. A device that puts the idea of free will into question may bring out these feelings in users.

Alternately, it may give users a feeling of control: if they know what their preferences are right now, they may be able to change them. After all, preferences change over time or given particular situations. It may also help users avoid choice paralysis (also called decision or analysis paralysis) by providing them with answers to certain choices they need to make throughout the day (e.g. Do I want a sandwich or a salad for lunch?).

Ultimately, it will be impossible to truly know how such a device might impact people until it is commercially available and adopted by a significant portion of the population. Still, it is fascinating to consider the potential implications.

The Human Factors Connection

Designing products, services, and environments requires a certain amount of knowledge about user preferences: will users prefer certain labeling? Will users prefer a push button or a dial? As we have mentioned before, context will be crucial when taking into account user preferences.

er one option over another, a:

These results indicate that by encouraging a user to repeatedly make decisions to choose a particular option, we can increase their preference for that option over time. When considering training programs or tasks that require repetitive action, we can keep this concept in mind as we apply psychology to our projects.

In a world of almost continuous technological advances, AI and machine learning are rapidly becoming commonplace. When working with machine learning and algorithms, it will be important to be mindful of potential racial bias. Some ways to mitigate bias in machine learning include:

- Matching the right model to the problem: there are many AI models available for use, so make sure that you take the time to consider which one is the best fit for what you are trying to do

- Use a representative training data set: try to get as close as you can to mirroring what the actual population looks like

- Monitor performance using real data: check your models and algorithms against real data as you develop them by using simple test questions

the World Economic Forum’s:

- Remember that individual differences matter: pay attention to how diversity affects each part of the process and endeavor to get a number of different perspectives at each step to ensure you consider different points of view

- Recognize that diversity in leadership is essential for dealing with the underrepresentation of minority voices: a diverse leadership team will be more likely to encourage and amplify the contributions of team members who might not otherwise be acknowledged

- Make sure that there is accountability: organizations need to ensure they have a high level of diversity in their leadership and employees to minimize the chances that bias will creep into the project at any point in the process

Finally, we will need to keep in mind that how users will react to a device that reads brain waves to tell them (and likely others) their preferences is still largely unknown. However, we can help keep it positive by providing ways for users to use this information to their benefit. This may be by incorporating it into planning technology to help with decision paralysis, prompting users to support them in making changes if they want to, and make design choices with the user’s mental comfort in mind.

There is a lot we can take from considering algorithms that can determine our preferences by reading our brain waves. Much of it can be applied right now: the science of how our preferences are formed and can change and how we can decrease algorithmic bias, or at least notice it and adjust for it. We have some exciting human factors challenges ahead of us; it will be interesting to see how these technological advances impact our lives in big and small ways.

For more Human Factors Cast content, check back every Tuesday for our news roundups and join us on Twitch every Thursday at 4:30 PM PST our weekly podcast. If you haven't already, join us on Slack and Discord or on any of our social media communities (LinkedIn, Facebook, Twitter, Instagram).

Follow

Chapters

Video

More from YouTube