How Social Media Is Adjusting To Ensure the Emotional Well-being of Its Users

As challenging as it is, social media platforms are starting to take steps to protect the emotional well-being of their users. Here are some of the ways they are doing so.
December 9, 2021
Share on facebook
Share on twitter
Share on linkedin

Get the best of
The Daily Bundle in your inbox every week

Get the best of The Daily Bundle in your inbox every week

Social media is under fire for a lot of things these days, from privacy matters to content ownership concerns. But maybe the biggest debate about social platforms is around mental health. There’s a lot of research about social media and its impact on its users, many of which found that these platforms can be highly detrimental to people’s mental health.

In that light, it’s not surprising to learn that people are demanding actions from social media platforms to protect themselves from bullying, gaslighting, depression, and isolation, among other issues. While tackling all those things can be hard, social media platforms have started stepping up to the challenge and are introducing new features and measures to ensure the emotional well-being of users.

Here are some of the most notable ones.

Expert-backed Resources

Probably the most common measure social media platforms have taken to help fight against mental issues is providing users with sets of expert resources filled with important information and actionable tips. Virtually all the major platforms have put forward this kind of resource hub, especially since the pandemic started.

That’s precisely what happened with Instagram, often pointed at as one of the most harmful social media platforms for its users. Though silent for a lot of time about the negative body image issues that plague the platform, Instagram launched a set of expert-backed resources for people with negative body image problems or eating disorders. This hub provides hotlines, advice, and in-depth information about these topics, popping up especially when users search for terms related to eating disorders or share content that addresses it. 

Facebook also introduced a set of mental health resources during this year, expanding on the already available contents of its Emotional Health Resource Center. The additions brought new global mental health guides and tips backed by the World Health Organization while also offering resources developed by different organizations across the world to help battle anxiety, depression, loss, and stress.

Pinterest had already introduced a similar mental resource hub in the U.S. back in 2019, which it later extended to other countries throughout 2020. The feature is called Compassionate Search, a name that speaks for itself about how this works: Once a user searches for terms and phrases of interest regarding stress, anxiety, depression, or difficult emotions, the feature pops up and offers well-being resources. 

TikTok and Snapchat both implemented their own set of resources as well. TikTok launched new features that automatically provide access to organizations that fight against negative body image issues and eating disorders. Snapchat, for its part, announced Here for You, a set of tools that promises proactive in-app support to users who may be having a mental health crisis. 

More Than Just Information

Naturally, users demand more than just access to information from other organizations. Given that the platforms are responsible for what is happening with their users, it’s only natural for people to expect stricter regulations and more control features that help prevent the most common issues in social media. Fortunately, platforms are introducing some interesting novelties. 

For instance, TikTok introduced a screen time management system into its users’ feeds back in 2020. The main goal of this system is to prompt a notification in users’ feeds whenever they’ve spent a lot of time scrolling through the app. This reminder is a very rare feature for social media platforms, which are always so keen to keep people on their sites. 

Snapchat also made improvements to its in-app reporting tools, making it easier for users to alert the platform whenever they feel like other users might be at risk of self-harm. This alert prompts the platform to contact the person and offer all the help available in Snapchat.

Regulations are a huge part of the deal when it comes to fighting mental health issues online. That’s why Instagram has implemented a stricter set of penalties for users sending abusive messages or using hate speech, which results in the deactivation of the accounts that send such messages. Instagram also provides users with the ability to better filter out comments and messages and manage DMs to prevent bullies or unwanted users from reaching out.

Both Facebook and Snapchat have taken a different approach to reach out to people with mental health issues. How? By developing their own video series targeted to audiences struggling with emotional distress. Facebook released Peace of Mind with Taraji, a Facebook Watch talk show hosted by Taraji P. Henson focusing on diverse mental health topics. Snapchat, for its part, did something similar with Mind Yourself, a docuseries hosted by Kevin Hart.

Finally, social media platforms are also partnering with mental health organizations to provide them with much-needed funding. Pinterest has committed to providing $10 million in funding for next year to help mental health organizations. The first beneficiary is the NPO #HalfTheStory, which raises awareness about the impact of social media on mental health. Twitter, the seemingly less active platform, has collaborated with the Cross-Government Working Group on Anti-Muslim Hatred to better understand how to fight hate speech on the platform. 

A Long Way to Go

It seems as if the pandemic has shaken the social media environment and pushed companies to do more about mental health issues. And while all the measures and features commented here are a nice start, they aren’t enough to tackle the widespread problems caused by social media. 

We still have a long way to go to really limit the impact of social media on mental health, so we need to remain active and keep demanding these changes from the platforms we use on a daily basis. That’s the only way we can make sure that social media evolves into a safe space for everyone to share and get together.

Notify of
Inline Feedbacks
View all comments

How useful was this post?

Click on a star to rate it!

Please enter a valid E-mail address.

Contact Us

How can we help you?

  • This field is for validation purposes and should be left unchanged.
Scroll to Top

Get in Touch

Jump-start your Business with the
Top 1% of IT Talent.

Need us to sign a non-disclosure agreement first? Please email us at [email protected].


By continuing to use this site, you agree to our cookie policy.