The days of having a dictionary on your bookshelf are numbered. But that’s OK, because everyone already walks around with a dictionary – not the one on your phone, but the one in your head.

Just like a physical dictionary, your mental dictionary[1] contains information about words. This includes the letters, sounds and meaning, or semantics, of words, as well as information about parts of speech and how you can fit words together to form grammatical sentences. Your mental dictionary is also like a thesaurus. It can help you connect words and see how they might be similar in meaning, sound or spelling.

Read more …Your mental dictionary is part of what makes you unique − here's how your brain stores and...

a person's hands resting on a laptop that's displaying a cartoon image of a head with an elongated nose that's poking through a representation of text on a page

Misinformation is a key global threat[1], but Democrats and Republicans disagree about how to address the problem. In particular, Democrats and Republicans diverge sharply on removing misinformation from social media.

Only three weeks after the Biden administration announced the Disinformation Governance Board in April 2022, the effort to develop best practices for countering disinformation was halted[2] because of Republican concerns about its mission. Why do Democrats and Republicans have such different attitudes about content moderation?

My colleagues Jennifer Pan[3] and Margaret E. Roberts[4] and I found in a study published in the journal Science Advances that Democrats and Republicans not only disagree about what is true or false, they also differ in their internalized preferences[5] for content moderation. Internalized preferences may be related to people’s moral values, identities or other psychological factors, or people internalizing the preferences of party elites.

And though people are sometimes strategic about wanting misinformation that counters their political views removed, internalized preferences are a much larger factor in the differing attitudes toward content moderation.

Internalized preferences or partisan bias?

In our study, we found that Democrats are about twice as likely as Republicans to want to remove misinformation, while Republicans are about twice as likely as Democrats to consider removal of misinformation as censorship. Democrats’ attitudes might depend somewhat on whether the content aligns with their own political views, but this seems to be due, at least in part, to different perceptions of accuracy.

Previous research showed that Democrats and Republicans have different views[6] about content moderation of misinformation. One of the most prominent explanations is the “fact gap”: the difference in what Democrats and Republicans believe is true or false. For example, a study found that both Democrats and Republicans were more likely to believe news headlines that were aligned with their own political views[7].

But it is unlikely that the fact gap alone can explain the huge differences in content moderation attitudes. That’s why we set out to study two other factors that might lead Democrats and Republicans to have different attitudes: preference gap and party promotion. A preference gap is a difference in internalized preferences about whether, and what, content should be removed. Party promotion is a person making content moderation decisions based on whether the content aligns with their partisan views.

We asked 1,120 U.S. survey respondents who identified as either Democrat or Republican about their opinions on a set of political headlines that we identified as misinformation based on a bipartisan fact check. Each respondent saw one headline that was aligned with their own political views and one headline that was misaligned. After each headline, the respondent answered whether they would want the social media company to remove the headline, whether they would consider it censorship if the social media platform removed the headline, whether they would report the headline as harmful, and how accurate the headline was.

Deep-seated differences

When we compared how Democrats and Republicans would deal with headlines overall, we found strong evidence for a preference gap. Overall, 69% of Democrats said misinformation headlines in our study should be removed, but only 34% of Republicans said the same; 49% of Democrats considered the misinformation headlines harmful, but only 27% of Republicans said the same; and 65% of Republicans considered headline removal to be censorship, but only 29% of Democrats said the same.

Even in cases where Democrats and Republicans agreed that the same headlines were inaccurate, Democrats were nearly twice as likely as Republicans to want to remove the content, while Republicans were nearly twice as likely as Democrats to consider removal censorship.

We didn’t test explicitly why Democrats and Republicans have such different internalized preferences, but there are at least two possible reasons. First, Democrats and Republicans might differ in factors like their moral values[8] or identities[9]. Second, Democrats and Republicans might internalize what the elites in their parties signal. For example, Republican elites have recently framed content moderation as a free speech[10] and censorship[11] issue. Republicans might use these elites’ preferences to inform their own.

When we zoomed in on headlines that are either aligned or misaligned for Democrats, we found a party promotion effect: Democrats were less favorable to content moderation when misinformation aligned with their own views. Democrats were 11% less likely to want the social media company to remove headlines that aligned with their own political views. They were 13% less likely to report headlines that aligned with their own views as harmful. We didn’t find a similar effect for Republicans.

Our study shows that party promotion may be partly due to different perceptions of accuracy of the headlines. When we looked only at Democrats who agreed with our statement that the headlines were false, the party promotion effect was reduced to 7%.

Implications for social media platforms

We find it encouraging that the effect of party promotion is much smaller than the effect of internalized preferences, especially when accounting for accuracy perceptions. However, given the huge partisan differences in content moderation preferences, we believe that social media companies should look beyond the fact gap when designing content moderation policies that aim for bipartisan support.

Future research could explore whether getting Democrats and Republicans to agree on moderation processes[12] – rather than moderation of individual pieces of content – could reduce disagreement. Also, other types of content moderation such as downweighting, which involves platforms reducing the virality of certain content, might prove to be less contentious. Finally, if the preference gap – the differences in deep-seated preferences between Democrats and Republicans – is rooted in value differences, platforms could try to use different moral framings[13] to appeal to people on both sides of the partisan divide.

For now, Democrats and Republicans are likely to continue to disagree over whether removing misinformation from social media improves public discourse or amounts to censorship.

Read more

Have you ever wondered whether the virus that gave you a nasty cold can catch one itself? It may comfort you to know that, yes, viruses can actually get sick. Even better, as karmic justice would have it, the culprits turn out to be other viruses.

Viruses can get sick in the sense that their normal function is impaired. When a virus enters a cell, it can either go dormant or start replicating right away[1]. When replicating, the virus essentially commandeers the molecular factory of the cell to make lots of copies of itself, then breaks out of the cell to set the new copies free.

Sometimes a virus enters a cell only to find that its new temporary dwelling is already home to another dormant virus. Surprise, surprise. What follows is a battle for control of the cell that can be won by either party.

But sometimes a virus will enter a cell to find a particularly nasty shock: a viral tenant waiting specifically to prey on the incoming virus.

I am a bioinformatician[2], and my laboratory[3] studies the evolution of viruses. We frequently run into “viruses of viruses,” but we recently discovered something new: a virus that latches onto the neck of another virus[4].

A world of satellites

Biologists have known of the existence of viruses that prey on other viruses – referred to as viral “satellites”[5] – for decades. In 1973, researchers studying bacteriophage P2, a virus that infects the gut bacterium Escherichia coli, found that this infection sometimes led to two different types of viruses emerging from the cell: phage P2 and phage P4[6].

Bacteriophage P4 is a temperate virus, meaning it can integrate into the chromosome of its host cell and lie dormant. When P2 infects a cell already harboring P4, the latent P4 quickly wakes up and uses the genetic instructions of P2[7] to make hundreds of its own small viral particles. The unsuspecting P2 is lucky to replicate a few times, if at all. In this case, biologists refer to P2 as a “helper” virus, because the satellite P4 needs P2’s genetic material to replicate and spread.

Bacteriophages are viruses that infect bacteria.

Subsequent research has shown that most bacterial species have a diverse set of satellite-helper systems[8], like that of P4-P2. But viral satellites are not limited to bacteria. Shortly after the largest known virus, mimivirus, was discovered in 2003, scientists also found its satellite, which they named Sputnik[9]. Plant viral satellites[10] that lurk in plant cells waiting for other viruses are also widespread and can have important effects on crops[11].

Viral arms race

Although researchers have found satellite-helper viral systems in pretty much every domain of life[12], their importance to biology remains underappreciated. Most obviously, viral satellites have a direct impact on their “helper” viruses, typically maiming them but sometimes making them more efficient killers[13]. Yet that is probably the least of their contributions to biology.

Satellites and their helpers are also engaged in an endless evolutionary arms race[14]. Satellites evolve new ways to exploit helpers and helpers evolve countermeasures to block them. Because both sides are viruses, the results of this internecine war necessarily include something of interest to people: antivirals.

Recent work indicates that many antiviral systems thought to have evolved in bacteria, like the CRISPR-Cas9 molecular scissors used in gene editing, may have originated in phages and their satellites[15]. Somewhat ironically, with their high turnover and mutation rates, helper viruses and their satellites turn out to be evolutionary hot spots for antiviral weaponry[16]. Trying to outsmart each other, satellite and helper viruses have come up with an unparalleled array of antiviral systems for researchers to exploit.

MindFlayer and MiniFlayer

Viral satellites have the potential to transform how researchers understand antiviral strategies, but there is still a lot to learn about them. In our recent work, my collaborators and I describe a satellite bacteriophage completely unlike previously known satellites, one that has evolved a unique, spooky lifestyle[17].

Undergraduate phage hunters[18] at the University of Maryland, Baltimore County isolated a satellite phage called MiniFlayer[19] from the soil bacterium Streptomyces scabiei. MiniFlayer was found in close association with a helper virus called bacteriophage MindFlayer[20] that infects the Streptomyces bacterium. But further research revealed that MiniFlayer was no ordinary satellite.

Microscopy image of a small round virus colored violet attached to the base of a larger round virus colored gray with a long tail
This image shows Streptomyces satellite phage MiniFlayer (purple) attached to the neck of its helper virus, Streptomyces phage MindFlayer (gray). Tagide deCarvalho[21]

MiniFlayer is the first satellite phage known to have lost its ability to lie dormant. Not being able to lie in wait for your helper to enter the cell poses an important challenge to a satellite phage. If you need another virus to replicate, how do you guarantee that it makes it into the cell around the same time you do?

MiniFlayer addressed this challenge with evolutionary aplomb and horror-movie creativity. Instead of lying in wait, MiniFlayer has gone on the offensive. Borrowing from both “Dracula” and “Alien,” this satellite phage evolved a short appendage[22] that allows it to latch onto its helper’s neck like a vampire. Together, the unwary helper and its passenger travel in search of a new host, where the viral drama will unfold again. We don’t yet know how MiniFlayer subdues its helper, or whether MindFlayer has evolved countermeasures.

If the recent pandemic has taught us anything, it is that our supply of antivirals is rather limited[23]. Research on the complex, intertwined and at times predatory nature of viruses and their satellites, like the ability of MiniFlayer to attach to its helper’s neck, has the potential to open new avenues for antiviral therapy.

Read more

A group of women in business dress sit around a conference table

On Oct. 4, 2022, the White House Office of Science and Technology Policy released the Blueprint for an AI Bill of Rights[1]: A Vision for Protecting Our Civil Rights in the Algorithmic Age. The blueprint launched a conversation about how artificial intelligence innovation can proceed under multiple fair principles. These include safe and effective systems, algorithmic discrimination protections, privacy and transparency.

A growing body of evidence highlights the civil and consumer rights that AI and automated decision-making jeopardize. Communities that have faced the most egregious discrimination historically now face complex and highly opaque forms of discrimination under AI systems. This discrimination occurs in employment, housing, voting, lending, criminal justice, social media, ad tech targeting, surveillance and profiling. For example, there have been cases of AI systems contributing to discrimination against women in hiring and racial discrimination[2] in the criminal justice system.

In the months that followed the blueprint’s release, the arrival of generative AI systems like ChatGPT added urgency to discussions about how best to govern emerging technologies in ways that mitigate risk without stifling innovation.

A year after the blueprint was unveiled, the Biden administration issued a broad executive order[3] on Oct. 30, 2023, titled Safe, Secure, and Trustworthy AI. While much of the order focuses on safety, it incorporates many of the principles in the blueprint.

The order includes several provisions that focus on civil rights and equity. For example, it requires that the federal government develop guidance for federal contractors on how to prevent AI algorithms from being used to exacerbate discrimination. It also calls for training on how best to approach the investigation and prosecution of civil rights violations related to AI and ensure AI fairness throughout the criminal justice system.

The vision laid out in the blueprint has been incorporated in the executive order as guidance for federal agencies. My research in technology and civil rights[4] underscores the importance of civil rights and equity principles in AI regulation.

Civil rights and AI

Civil rights laws often take decades or even lifetimes to advance. Artificial intelligence technology and algorithmic systems are rapidly introducing black box[5] harms such as automated decision-making that may lead to disparate impacts. These include racial bias in facial recognition systems.

These harms are often difficult to challenge, and current civil rights laws and regulations may not be able to address them. This raises the question of how to ensure that civil rights are not compromised as new AI technologies permeate society.

When combating algorithmic discrimination, what does an arc that bends toward justice look like? What does a “Letter from Birmingham Jail[6]” look like when a civil rights activist is protesting not unfair physical detention but digital constraints such as disparate harms from digitized forms of profiling, targeting and surveillance?

The 2022 blueprint was developed under the leadership of Alondra Nelson[7], then acting director[8] of the Office of Science and Technology Policy[9], and her team. The blueprint lays out a series of fair principles that attempt to limit a constellation of harms that AI and automated systems can cause.

Beyond that, the blueprint links the concepts of AI fair principles and AI equity to the U.S. Constitution and the Bill of Rights. By associating these fair principles with civil rights and the Bill of Rights, the dialogue can transition away from a discussion that focuses only on a series of technical commitments, such as making AI systems more transparent. Instead, the discussion can address how the absence of these principles might threaten democracy.

Arati Prabhakar, director of the White House Office of Science and Technology Policy, and Alondra Nelson, former acting director, discussed the Blueprint for an AI Bill of Rights at a conference on the anniversary of its release.

A few months after the release of the blueprint, the U.S. Department of Civil Rights Division, the Consumer Financial Protection Bureau, the Equal Employment Opportunity Commission and the Federal Trade Commission jointly pledged to uphold the U.S.’s commitment[10] to the core principles of fairness, equality and justice as emerging automated systems become increasingly common in daily life. Federal[11] and state legislation[12] has been proposed to combat the discriminatory impact of AI and automated decision-making.

Civil rights organizations take on tech

Multiple civil rights organizations, including the Leadership Conference on Civil and Human Rights[13], have made AI-based discrimination a priority. On Sept. 7, 2023, the Leadership Conference launched[14] a new Center for Civil Rights and Technology[15] and tapped Nelson, author of the Blueprint for an AI Bill of Rights, as an adviser.

Before the release of the new executive order, Sen. Ed Markey, Rep. Pramila Jayapal and other members of Congress sent a letter to the White House urging the administration to incorporate the blueprint’s principles[16] into the anticipated executive order. They said that “the federal government’s commitment to the AI Bill of Rights would show that fundamental rights will not take a back seat in the AI era.”

Numerous civil rights and civil society organizations sent a similar letter to the White House[17], urging the administration to take action on the blueprint’s principles in the executive order.

As the Blueprint for an AI Bill of Rights passed its first anniversary, its long-term impact was unknown. But, true to its title, it presented a vision for protecting civil rights in the algorithmic age. That vision has now been incorporated in the Executive Order on Safe, Secure, and Trustworthy AI. The order can’t be properly understood without this civil rights context.

Read more

a woman speaks at a podium to a room of seated people while a man stands nearby watching

The comprehensive, even sweeping, set of guidelines[1] for artificial intelligence that the White House unveiled in an executive order on Oct. 30, 2023, show that the U.S. government is attempting to address the risks posed by AI.

As a researcher of information systems and responsible AI[2], I believe the executive order represents an important step in building responsible[3] and trustworthy[4] AI.

The order is only a step, however, and it leaves unresolved the issue of comprehensive data privacy legislation. Without such laws, people are at greater risk of AI systems revealing sensitive or confidential information[5].

Understanding AI risks

Technology is typically evaluated for performance, cost and quality[6], but often not equity, fairness and transparency. In response, researchers and practitioners of responsible AI have been advocating for:

The National Institute of Standards and Technology (NIST) issued a comprehensive AI risk management framework[13] in January 2023 that aims to address many of these issues. The framework serves as the foundation[14] for much of the Biden administration’s executive order. The executive order also empowers the Department of Commerce[15], NIST’s home in the federal government, to play a key role in implementing the proposed directives.

Researchers of AI ethics have long cautioned that stronger auditing of AI systems[16] is needed to avoid giving the appearance of scrutiny without genuine accountability[17]. As it stands, a recent study looking at public disclosures from companies found that claims of AI ethics practices outpace actual AI ethics initiatives[18]. The executive order could help by specifying avenues for enforcing accountability.

Another important initiative outlined in the executive order is probing for vulnerabilities of very large-scale general-purpose AI models[19] trained on massive amounts of data, such as the models that power OpenAI’s ChatGPT or DALL-E. The order requires companies that build large AI systems with the potential to affect national security, public health or the economy to perform red teaming[20] and report the results to the government. Red teaming is using manual or automated methods to attempt to force an AI model to produce harmful output[21] – for example, make offensive or dangerous statements like advice on how to sell drugs.

Reporting to the government is important given that a recent study found most of the companies that make these large-scale AI systems lacking[22] when it comes to transparency.

Similarly, the public is at risk of being fooled by AI-generated content. To address this, the executive order directs the Department of Commerce to develop guidance for labeling AI-generated content[23]. Federal agencies will be required to use AI watermarking[24] – technology that marks content as AI-generated to reduce fraud and misinformation – though it’s not required for the private sector.

The executive order also recognizes that AI systems can pose unacceptable risks[25] of harm to civil and human rights[26] and the well-being of individuals: “Artificial Intelligence systems deployed irresponsibly have reproduced and intensified existing inequities, caused new types of harmful discrimination, and exacerbated online and physical harms.”

The U.S. government takes steps to address the risks posed by AI.

What the executive order doesn’t do

A key challenge for AI regulation is the absence of comprehensive federal data protection and privacy legislation. The executive order only calls on Congress to adopt privacy legislation, but it does not provide a legislative framework. It remains to be seen how the courts will interpret the executive order’s directives in light of existing consumer privacy and data rights statutes.

Without strong data privacy laws in the U.S. as other countries have, the executive order could have minimal effect on getting AI companies to boost data privacy. In general, it’s difficult to measure the impact that decision-making AI systems have on data privacy and freedoms[27].

It’s also worth noting that algorithmic transparency is not a panacea. For example, the European Union’s General Data Protection Regulation legislation mandates “meaningful information about the logic involved[28]” in automated decisions. This suggests a right to an explanation of the criteria that algorithms use in their decision-making. The mandate treats the process of algorithmic decision-making as something akin to a recipe book, meaning it assumes that if people understand how algorithmic decision-making works, they can understand how the system affects them[29]. But knowing how an AI system works doesn’t necessarily tell you why it made a particular decision[30].

With algorithmic decision-making becoming pervasive, the White House executive order and the international summit on AI safety[31] highlight that lawmakers are beginning to understand the importance of AI regulation, even if comprehensive legislation is lacking.

Read more

The cars, cellphones, computers and televisions that people in the U.S. use every day require metals like copper, cobalt and platinum to build. Demand from the electronics industry for these metals is only rising, and companies are constantly searching for new places on Earth to mine them.

Scientists estimate that lots of these metals exist thousands of miles beneath Earth’s surface, in its molten core, but that’s far too deep and hot to mine[1]. Instead, some companies hope to one day search for deposits that are literally out of this world — on asteroids.

The commercialization of asteroid mining is still a ways off, but in October 2023, NASA launched a scientific mission to explore the metal-rich asteroid Psyche[2]. The main goal of the mission[3] is studying the composition and structure of this asteroid, which could tell scientists more about Earth’s core since the two objects might have a similar makeup.

Both likely contain platinum, nickel, iron and possibly even gold – materials of commercial interest.

Experts need to know what’s out there on asteroids before considering whether they’re worth mining. NASA’s Psyche mission could answer some of these questions.

I am a planetary geologist[4] whose work explores other planets and astronomical objects like Mars, Venus and the Moon. I will be following the Psyche mission closely, as this is the first time that scientists will be able to learn about the composition and structure of a possible piece of a planetary core similar to the Earth’s[5], without indirect seismic or magnetic measurements, or replicating the pressure and temperature conditions of the Earth’s core in our labs.

With the spacecraft estimated to arrive at the asteroid’s orbit in 2029, the findings from the Psyche mission will provide unique insights into the type of metals present on the asteroid’s surface, as well as their amount, and the minerals containing these metals. This data is essential both for scientists like me exploring the formation and evolution planetary bodies, as well as for companies investigating the possibility of asteroid mining.

Asteroid formation

Asteroids come in a variety of sizes[6]. Some are the size of a town, while others are the size of a state. Most asteroids are made of rocks and represent the leftovers from the early formation of our solar system[7] around 4.6 billion years ago.

An artist's illustration of a gray asteroid with some yellow-ish surfaces, and two large circular craters.
The Psyche asteroid. NASA/JPL-Caltech/ASU[8]

Not every asteroid is the same – some, like Bennu, the target of NASA’s OSIRIS-REx mission[9], are rich in carbon. These are very old, and they will teach scientists more about how planets formed and how life may have begun on Earth.

Others, like Psyche[10], are made of metals and potentially result from one or more collisions between astronomical objects when the solar system was forming. These collisions left debris flying through space — including potential pieces of a planet’s metal-rich core. A NASA spacecraft will orbit and analyze the surface of Psyche.

Mining in space

Not every mineral deposit on Earth is mineable. Companies first look for deposits with a high level of metal purity[11]. They also investigate how affordable and feasible extracting the metal would be before choosing where to mine.

A bird's eye view of a gray rock, with a red crane on it.
Before mining, companies think about whether a deposit will yield enough metal. The same principle applies to asteroid mining. Abstract Aerial Art/DigitalVision via Getty Images[12]

Similarly, before mining an asteroid, companies will have to think about all those factors, and they’ll have to come up with the infrastructure needed to mine at a distance and transport the metals they mine hundreds of millions of miles back to Earth. The technology to do that is still years away, and transporting metals would require major funding.

A few companies[13] around the world have already started to think about what the best and lowest cost approach would be, drawing from processes similar to those used on Earth.

The first step would be finding a mineable metal deposit[14]. Next, they’d drill and extract the metals on the asteroid[15]. One of the most important differences with Earth mines is that each step would be undertaken remotely with spacecrafts orbiting around the asteroid and robots landing on its surface. Then, a spacecraft would send the resulting materials back to Earth.

Asteroid mining plans are still at their earliest stages. A few companies like Planetary Resources[16] and Deep Space Industries[17], with goals to extract metals from space, were acquired by other companies.

Experts can’t quite tell yet how acquiring valuable metals from asteroids would affect the global economy, but these metals could potentially flood the market and lower their values[18].

The Psyche mission is a huge step in figuring out what sort of metals are out there, and it may also answer questions about the composition and properties of Earth’s core.

Read more

The benefits of friendship go far beyond having someone to confide in or spend time with – it can also protect you from physical and mental health problems. For example, people with good friends recover more quickly from illnesses[1] and surgeries[2]. They report higher well-being[3] and feel like they live up to their full potential[4]. Additionally, people with good friends report being less lonely across many life stages, including adolescence[5], becoming a parent[6] and old age[7].

In fact, friendships are so powerful that the social pain of rejection activates the same neural pathways that physical pain[8] does.

Behavioral scientists like me[9] have tended to focus our research about friendships on their benefits. How to cultivate these powerful relationships hasn’t been as deeply researched yet. Understanding more about what people look for in a friend and how to make and sustain good friendships could help fight the loneliness epidemic[10].

Traditional conceptions of friendship

Previous generations of behavioral scientists traditionally focused on the notion that people form friendships with those who are similar[11], familiar[12] and in close proximity[13] to them.

When you look at all the friendships you’ve had over your life, these three factors probably make intuitive sense. You’re more likely to have things in common with your friends than not. You feel an increased sense of familiarity with friends the longer you know them – what psychologists call the mere exposure effect[14]. And your friends are more likely to live or work near you.

Researchers in this field have also typically divided friendship preferences based on gender. The dichotomy suggests that women prefer one-on-one[15], emotionally close[16] and face-to-face[17] friendships, while men prefer multi-person, task-oriented and side-by-side[18] friendships, with the focus on a shared activity.

two seated women laughing with mugs in their hands
Research suggests that women on average prefer a one-on-one, close friendship style. FG Trade/E+ via Getty Images[19]

Again, when looking at your own friendships, these findings may seem intuitive. Women on average prefer to engage in activities that allow for self-disclosure and sharing secrets, such as spending time one-on-one talking about their lives. Men, on the other hand, tend to prefer to engage in activities that are group-based and have a clearly defined outcome, such as playing sports together. Findings such as these show that gender[20] and preferences on how to connect[21] are important in friendships.

But these explanations of friendship do not address the most important aspect of making friends – choosing the individual people you want to turn into your pals. Friendship decisions are not random. There are many people who are similar, familiar, in close proximity and have similar preferences as you. Yet few of these individuals end up being your friends.

So, in a world full of possibilities, how do people pick those who will become their friends?

New ways to think about friendship

Within the last decade, researchers have begun investigating the roots of friendship preferences beyond the classic descriptions.

For example, social scientists see there are strong preferences for friends to be loyal, trustworthy[22] and warm[23]. Additionally, researchers find there are preferences for friends who help you solve specific kinds of problems[24] and are generous and caring with you[25] instead of others. These preferences help people navigate making friends, given limited reserves of time and effort[26]. In short, they help you find the best possible friends you can in a world full of friendship possibilities.

Social scientists have also learned that, while there are some important gender differences in what people want in friends, it is not accurate to say that men and women want one kind of friendship over another. In fact, when we take a more holistic approach and consider[27] broader categorizations of[28] emotional closeness[29] and tasks[30], the gender differences in these preferences are reduced. And of course, people don’t exclusively pick between face-to-face and side-by-side friendships. Instead, it is more likely that they focus on what they want from their friends and let these needs guide how friendships form.

Ultimately it’s your individual preferences that guide you toward the people who will best meet your particular social needs. With a little luck, you’ll find buddies who can lend a hand when you need one and support you in reaching your goals. In all, your preferences are the key to finding friends who can buffer against feeling lonely and provide you with the social, emotional and health benefits of friendship.

smiling man in bike helmet in foreground of a bike group pit stop
Knowing the kind of friendship you prefer can help you figure out where to look for friend possibilities. Thomas Barwick/DigitalVision via Getty Images[31]

When you’re looking for friends

It’s hard to provide clear guidelines for improving friendships because the research about friendship preferences is still developing. But there are some clear points for consideration:

  1. Determine what you value in friends. Do you want one-on-one, emotionally close friendships or multi-person, task-oriented friendships? Depending on your preference, different kinds of activities will be helpful for finding others who fit the bill and cultivating these friendships.

  2. Know that it will take time to make close friendships[32]. Research suggests that it takes 30 hours of interaction to make a casual friend, 140 hours to make a good friend and 300 hours to make a best friend.

  3. Consider what you bring to the table. Everyone has unique strengths they bring to their friendships. Research shows that, when you’re able to demonstrate that you have characteristics people want in friends, you’re able to make more satisfying friendships[33].

Understand friendships to understand loneliness

Considering the nuances of friendship preferences will be extremely important in reducing not only loneliness, but other related public health crises. For example, loneliness is associated with likelihood of attempting suicide[34]. Recent surveys have found that men are suffering big declines in the number of close friends[35] they have, as well as experiencing higher rates of suicide[36] compared to women.

The U.S. Surgeon General’s recent recommendations for fighting the loneliness epidemic focus on public policies and infrastructure[37]. But fostering community spaces for connection – such as parks, libraries and playgrounds – prioritizes the preferences of those who favor the one-on-one, emotionally close and face-to-face connections more often preferred by women. These places are less beneficial for people with more typically masculine preferences, as there is no guarantee that these spaces will foster side-by-side, task-oriented connections unless areas for sports and other team-based activities are also included.

To counter this inequity, researchers and public health officials first need to understand what makes friendships satisfying. Then they can ensure that recommendations to curb loneliness address all of the pathways that people use to cultivate high-quality friendships.

Read more

More Articles …