FACEBOOK, INC. | Report on Community Standards Enforcement

AGM date
Proposal number
Resolution details
Company ticker
Lead filer
Resolution ask
Report on or disclose
ESG theme
  • Social
ESG sub-theme
  • Health, safety and well being
Company sector
Company HQ country
United States
Resolved clause
Shareholders request the Board, at reasonable expense and excluding proprietary or legally privileged information, prepare a report analyzing why
the enforcement of “Community Standards” as described in the “Transparency Center” has proven ineffective at controlling the dissemination of user content that
contains or promotes hate speech, disinformation, or content that incites violence and/or harm to public health or personal safety.
Whereas clause
The Meta (formerly Facebook) brand has continued to be wracked by management missteps and lack of Board oversight, resulting in continued
harm by its platform including:
• Millions of high-profile users exempted from its rules, permitting continued widespread; incitement of violence and harrassment;
• Internal Company research demonstrating that Instagram is toxic for teen girls;
• Mental health crises among outsourced moderators due to viewing child pornography and animal cruelty;
• Lack of cooperation with authorities to prevent and detect child exploitation and abuse;
• Ignored employee red flags about the spread of election misinformation;
• Political advertisements containing deliberate lies and mistruths;
• Hate speech that continues to thrive;
• Anti-immigrant violence around the world.
A whistleblower complaint filed with the SEC argues that the Company has failed to adequately warn investors about the material risks of dangerous and
criminal behavior, terrorist content, hate speech, and misinformation on its sites. Company failure to control these activities reflects a grave lack of oversight by
management and the board. Despite establishing an internal Oversight Board, the Company’s platforms continue to harm society and create investor risk. An
internal review of company practices highlighting harassment and incitement to violence states, “We are not actually doing what we say we do publicly,” and
deems company’s actions “a breach of trust.”
Management has attempted to address the material risk of dangerous user content through the creation of the “Transparency Center” that displays qualitative
and quantitative reports on the elimination of posts that violate the 25 “Community Standards.” Shareholders applaud this action, yet ask why this seemingly
robust technological and human-screening system is ineffective?
Supporting statement
Proponent suggests the report include, in Board and management discretion:
• A quantitative and qualitative assessment by an external, independent panel of qualified computer scientists of the effectiveness of Meta’s algorithms to
locate and eliminate content that violates the Community Standards
• An assessment of the effectiveness of Meta’s staff and contractors in locating and eliminating content that violates the Community Standards
• An examination of benefits to users and impact to revenue if the Company would voluntarily follow existing legal frameworks established for broadcast
networks (e.g. laws forbidding child pornography and rules governing political ads)
• An analysis of the benefits of the Company continuing to conduct technology impact assessments focused on how Meta’s platforms affect society.
This report

How other organisations have declared their voting intentions

Organisation name Declared voting intentions Rationale
Universities Superannuation Scheme - USS For