As X grapples with content moderation criticism, Meta, the parent company of Facebook and Instagram, faces its own queries regarding the accuracy of its reported moderation statistics and the protection of young users.
A federal lawsuit, representing 33 states, claims that Meta has allegedly misrepresented its Community Standards Enforcement Reports. These reports showcase low rates of community standards violations while excluding key data from user experience surveys. Internal surveys allegedly contradict the reported rates, suggesting a much higher incidence of harmful content exposure for users.
The complaint asserts that Meta used a skewed data approach, presenting lower averages by dividing a smaller number of reports across its vast user base. However, internal user feedback suggests significantly higher rates of encountering harmful content, indicating a discrepancy between reported figures and actual user experiences.
Moreover, the lawsuit reveals alarming figures of more than 1.1 million reports of users under 13 accessing Instagram since early 2019, with only a fraction of these accounts reportedly disabled by Meta.
These allegations, filed in the U.S. District Court for the Northern District of California, highlight potential violations of privacy laws. If proven, Meta could face substantial fines and intensified scrutiny over its moderation and safety measures, particularly concerning younger users' access.
Meta has countered these allegations, stating that the complaint misrepresents its efforts by utilizing selective quotes and cherry-picked documents. The lawsuit's outcome may shed light on the accuracy of Meta's reported statistics and its commitment to user safety.