Social media giant Meta said that over 19.3 million content pieces were "actioned" on Facebook across 13 violation categories proactively across India in December 2021.
Its photo-sharing platform Instagram took action against over 2.4 million pieces across 12 categories during the same period proactively, according to data shared in a compliance report.
Under the IT rules that came into effect in May last year, large digital platforms (with over five million users) have to publish periodic compliance reports every month, mentioning the details of complaints received and action taken thereon. It also includes details of content removed or disabled via proactive monitoring using automated tools.
Facebook had "actioned" over 16.2 million content pieces proactively in October across 13 categories, while Instagram took action against over 3.2 million pieces across 12 categories during the same period proactively.
"Over the years, we have consistently invested in technology, people and processes to further our agenda of keeping our users safe and secure online and enabling them to express themselves freely on our platform. We use a combination of artificial intelligence, reports from our community and review by our teams to identify and review content against our policies," a Meta spokesperson said.
The spokesperson added that in accordance with the IT rules, the company has published its monthly compliance report for the period of December 1 to December 31, and will contain details of the content that it has removed proactively using automated tools and details of user complaints received and action taken.
In its latest report, Meta said 531 user reports were received by Facebook through its Indian grievance mechanism during December 1-31, 2021.
"Of these incoming reports, we provided tools for users to resolve their issues in 436 cases," the report said.
These include pre-established channels to report content for specific violations, self-remediation flows where they can download their data, avenues to address account hacked issues, etc, it added.
During December 1 to December 31 period, Instagram received 436 reports through the Indian grievance mechanism.
Facebook's parent company recently changed its name to Meta. Apps under Meta include Facebook, WhatsApp, Instagram, Messenger and Oculus.
As per the latest report, the over 19.3 million content pieces actioned by Facebook during December included content related to spam (13.8 million), violent and graphic content (2.1 million), adult nudity and sexual activity (1.5 million), and hate speech (60,800).
Other categories under which content was actioned include bullying and harassment (1,17,000), suicide and self-injury (3,74,300), dangerous organisations and individuals: terrorist propaganda (1,18,300) and dangerous organisations and Individuals: organised hate (20,000).
Categories like child endangerment- nudity and physical abuse category saw 1,57,100 content pieces being actioned, while child endangerment- sexual exploitation saw 7,96,800 pieces, and under violence and incitement, 2,67,100 pieces were actioned.
"Actioned" content refers to the number of pieces of content (such as posts, photos, videos or comments) where action has been taken for violation of standards. Taking action could include removing a piece of content from Facebook or Instagram or covering photos or videos that may be disturbing to some audiences with a warning.
The proactive rate, which indicates the percentage of all content or accounts acted on which Facebook found and flagged using technology before users reported them, in most of these cases ranged between 57.4-99.9 per cent.
The proactive rate for removal of content related to bullying and harassment was 57.4 per cent as this content is contextual and highly personal by nature. In many instances, people need to report this behaviour to Facebook before they can identify or remove such content.
For Instagram, over 2.4 million pieces of content were actioned across 12 categories during December 2021. This includes content related to suicide and self-injury (8,91,900), violent and graphic content (6,00,800), adult nudity and sexual activity (4,61,900), and bullying and harassment (2,09,200).
Other categories under which content was actioned include hate speech (16,700), dangerous organisations and individuals: terrorist propaganda (10,700), dangerous organisations and individuals: organised hate (2,500), child endangerment- nudity and physical abuse (21,500), child endangerment- sexual exploitation (1,71,200) and violence and incitement (25,600).