Difference between revisions of "Facebook and Disinformation"

From BigTechWiki
Jump to navigationJump to search
Line 1: Line 1:
* Laura Edelson Said Her And Her Team Used Their Research To study Facebook’s “apparent amplification of partisan misinformation.” Edelson and her team found that Facebook users engaged with misinformation more than other kinds of information on the platform. The team’s findings led Edelson to believe we were “racing against the clock” to understand how disinformation spread on social media. Edelson called understanding how misinformation spread on social media a “right now” problem.
* Laura Edelson said she and her team used their research to study Facebook’s “apparent amplification of partisan misinformation.” Edelson and her team found that Facebook users engaged with misinformation more than other kinds of information on the platform. The team’s findings led Edelson to believe we were “racing against the clock” to understand how disinformation spread on social media. Edelson called understanding how misinformation spread on social media a “right now” problem.
** Facebook’s XCheck allowed whitelisted users to post inflammatory claims even when they had been deemed false by Facebook’s fact checkers. Misleading posts by whitelisted users said that vaccines were deadly, that Hillary Clinton had covered up pedophile rings and that Trump called asylum seekers “animals.”<ref>https://www.wsj.com/articles/facebook-files-xcheck-zuckerberg-elite-rules-11631541353</ref>
** Facebook’s XCheck allowed whitelisted users to post inflammatory claims even when they had been deemed false by Facebook’s fact checkers. Misleading posts by whitelisted users said that vaccines were deadly, that Hillary Clinton had covered up pedophile rings and that Trump called asylum seekers “animals.”<ref>https://www.wsj.com/articles/facebook-files-xcheck-zuckerberg-elite-rules-11631541353</ref>
** In 2020, posts by whitelisted users that contained misinformation had been viewed at least 16.4 billion times.
** In 2020, posts by whitelisted users that contained misinformation had been viewed at least 16.4 billion times.
** Facebook itself was a proliferator of misinformation. An internal review of Facebook’s whitelisting practice said “we are not actually doing what we say we do publicly.” Facebook even lied to its own oversight board about XCheck and the whitelisting of users, saying the system was used in “a small number of decisions.”
** Facebook itself was a proliferator of misinformation. An internal review of Facebook’s whitelisting practice said “we are not actually doing what we say we do publicly.” Facebook even lied to its own oversight board about XCheck and the whitelisting of users, saying the system was used in “a small number of decisions.”
** Zuckerberg opposed reforming Facebook’s algorithm to stop it from rewarding misinformation over business concerns and fears it would hurt efforts to increase engagement by users.
** Zuckerberg opposed reforming Facebook’s algorithm to stop it from rewarding misinformation over business concerns and fears it would hurt efforts to increase engagement by users.

Revision as of 18:44, 23 March 2022

  • Laura Edelson said she and her team used their research to study Facebook’s “apparent amplification of partisan misinformation.” Edelson and her team found that Facebook users engaged with misinformation more than other kinds of information on the platform. The team’s findings led Edelson to believe we were “racing against the clock” to understand how disinformation spread on social media. Edelson called understanding how misinformation spread on social media a “right now” problem.
    • Facebook’s XCheck allowed whitelisted users to post inflammatory claims even when they had been deemed false by Facebook’s fact checkers. Misleading posts by whitelisted users said that vaccines were deadly, that Hillary Clinton had covered up pedophile rings and that Trump called asylum seekers “animals.”[1]
    • In 2020, posts by whitelisted users that contained misinformation had been viewed at least 16.4 billion times.
    • Facebook itself was a proliferator of misinformation. An internal review of Facebook’s whitelisting practice said “we are not actually doing what we say we do publicly.” Facebook even lied to its own oversight board about XCheck and the whitelisting of users, saying the system was used in “a small number of decisions.”
    • Zuckerberg opposed reforming Facebook’s algorithm to stop it from rewarding misinformation over business concerns and fears it would hurt efforts to increase engagement by users.