The announcement comes after major advertisers pulled their ads from the video platform following a viral video that showed comments on some videos sexualized minors.
"We recognize that comments are a core part of the YouTube experience and how you connect with and grow your audience," the company wrote in a blog post. "At the same time, the important steps we're sharing today are critical for keeping young people safe."
The company said that in the past week it has taken "a number of steps to better protect children and families" by suspending comments that violate its policies and deleting "tens of millions of videos," but now will take its efforts one step further and begin to suspend comments on "most" videos that feature minors.
The announcement comes after the video platform faced a backlash from advertisers, including AT&T, Nestlé and Epic Games, following a viral video that claimed YouTube-hosted videos of young children were found to attract comments from apparent pedophiles.
"Until Google can protect our brand from offensive content of any kind, we are removing all advertising from YouTube," an AT&T spokesperson said in an email to NBC News on Feb. 21.
YouTube said there will be a small number of creators that will be able to keep their comments enabled on their videos, but those accounts will have to "actively monitor" their comment feeds and demonstrate "low predatory behavior."
In addition to suspending comments, the video platform said they will be launching a comments "classifier" that will be able to identify and remove "predatory" comments much quicker.
YouTube's announcement also comes a day after the Federal Trade Commission fined popular music app TikTok $5.7 million on allegations of illegally collecting images, voice records and geolocations of children, some younger than 13.