Quantcast
Channel: Web Science and Digital Libraries Research Group
Viewing all articles
Browse latest Browse all 738

2024-06-12: Paper Summary: Exploring the Use of Personalized AI for Identifying Misinformation on Social Media

$
0
0


Figure 1 Jahanbakhsh et al. Assess tweets with AI assistance (UI used in Step 3). Participants assess each tweet in their feed while seeing AI model predictions. As shown in this screenshot, the user is in the middle of an experiment and has assessed the top tweet, while the rest of the tweets in this feed (for which the AI model predictions are shown) have not been assessed yet. The AI model predictions may change in response to user feedback. The tweets for which the AI predictions have recently changed are shown on the left pane. Newly changed predictions are differentiated visually with a border and a notification icon, similar to the bottom tweet in the image (Figure 3 in original paper).

In recent years, misinformation on social media platforms has been a tremendous concern. The spread of inaccurate or misleading information weakens the credibility of online content and also poses serious threats to social trust, public opinion, and even democratic processes. Researchers and technologists have been exploring various approaches to prevent misinformation effectively. One promising method is the use of personalized Artificial Intelligence (AI) systems capable of identifying and preventing the spread of misinformation customized to individual users. These AI systems pose advanced algorithms and user-specific data to deliver personalized content recommendations, fact-checking, and contextual information, aiming to empower users to make more informed decisions when engaging with online content.

In "Exploring the Use of Personalized AI for Identifying Misinformation on Social Media" published in CHI 2023, the authors, Farnaz Jahanbakhsh, Yannis Katsis, Dakuo WangLucian Popa, Michael Muller have explored the potential of personalized AI systems in identifying misinformation on social media platforms. Personalized AIs can detect and counteract misinformation, considering factors such as user behavior, preferences, and past engagement with online content. They also investigate the ethical implications and challenges associated with the deployment of personalized AI systems for misinformation detection. The authors contribute to a deeper understanding of how AI can be utilized to promote a more trustworthy and informed online environment.


Understanding Personalized AI


Personalized AIs refer to algorithms customized to individual users based on their preferences, behaviors, and past interactions. Social media platforms use personalized AI to curate users' newsfeeds and also show them content related to their interests.


The Role of Personalized AI in Misinformation Detection


By analyzing the user's past interactions, preferences and the credibility of the sources they typically engage with, personalized AI algorithms can asses the trustworthiness of the information they encounter on social media. For example, if a user typically engages with sources known for their accuracy and suddenly interacts with information from a source with a questionable reputation, the AI algorithm can flag it as potentially misinformation. Similarly, if a user frequently shares misinformation, the algorithm can adjust the content they see to include more credible sources.


Methodology


The researchers developed a platform enabling human-AI interaction. Participants assessed  feed of tweets and receive predictions from their personalized AI regarding how they would assess other tweets. As participants provided more assessments, their personalized AI would retrain and update its predictions. The task served two main purposes:

  1. Exposing participants to interacting with a personalized AI to determine content accuracy on social media and gauge their perceptions.
  2. Understanding whether a user's assessment of a tweet would be influenced by an AI predicting how the user might assess it.

To study the AI's influence on user decisions, they couldn't ask users to assess tweet accuracy before and after seeing AI predictions, as this might bias their assessments. Instead, they compared user agreement with the AI across tweets, some with AI predictions displayed and some without. To prevent confounding effects, they trained two separate AI models evolving independently. This allowed the researchers to display one model's predictions on some tweets and withhold the predictions of the other AI on different tweets. By comparing how often users agreed with the predictions, they ensured that any observed differences were attributed to users seeing or not seeing AI predictions, rather than differences in AI model performance.


Data Collection


The dataset includes user interactions and content shared on Twitter, a popular social media platform. It also has user profiles, posts, comments, likes, shares, and the credibility scores of the sources. Additional information on user engagement patterns, such as the types of content users typically interacted with and the frequency of their interactions. Table 1 shows sample topics from collected data. 


The tweets were collected based on the following criteria:

  • Tweets must contain verifiable claims, excluding arguments, opinions, or life stories.
  • Tweets should be in the form of statements or evident rhetorical questions, not questions, to ensure binary accuracy labeling.
  • The topic of tweets should not relate to news outside the United States to maintain relevance for US-based users.
  • Tweets must be self-contained, requiring no interpretation from other tweets to understand.
  • The main topic of the tweet should be COVID-19, excluding topics like the pandemic's economic impact to limit subtopics.
  • Tweets should not focus on COVID-19 statistics like death tolls, as these numbers may change before the user study begins.

 Table 1 Jahanbakhsh et al. - The two sides of the arguments in COVID-19 related subtopics from pool of tweets. (Table 1 in original paper)

Preprocessing


Before training the personalized AI model, the authors preprocessed the collected data. This included cleaning the text data, removing stop words, tokenizing, and applying vectorization techniques such as TF-IDF (Term Frequency-Inverse Document Frequency) and word embeddings.

  • Tweets for both AI processing and participant presentation were cleaned.
  • Removed @username strings at the beginning of tweets, replacing other occurrences within the text with a special token.
  • Replaced links with another special token.
  • Removed hashtags but retained the word following the hashtag.
  • Eliminated indicators of tweet threads at the end of the tweet text (e.g., 2/5).
  • Only the tweet texts were used for training users' personalized AI models; other features such as tweet author or date were not considered due to their rarity in the dataset.

Training Data


The dataset divides into training, validation, and test sets. The training data included user interactions with both credible and non-credible sources of information. They labeled the data with user interactions as either reliable or unreliable, based on the credibility scores of the sources.


Personalized AI Model


They developed a personalized AI model using machine learning algorithms such as Support Vector Machines (SVM), Random Forest, and Neural Networks. The model trained to predict the credibility of information based on a user's past interactions and preferences. To train a personalized model for each user that evolved as the user provided more assessments, the researchers required an AI system that could be quickly retrained. They deployed an instance of an open-source system called Label Sleuth for this purpose. Label Sleuth is an interactive visual system that allows users to label text elements while the system automatically trains in the background. They also customized Label Sleuth to communicate with the experimental system. Continuously submit a user's assessments, check for updates to the user's model, and retrieve the most recent model's predictions using Label Sleuth's backend API. Label Sleuth trains a binary classifier that predicts which items from a document (tweet texts) likely belong to the positive category, which, in their case, was the category "inaccurate". The model training happens iteratively, starting a new model training iteration as more labels are provided.


To allow for fine-grained control over when a new iteration is invoked, Label Sleuth offers two customizable settings: positive threshold and changed element threshold. These settings denote after how many initial positive labels a model should start training for the user and the number of changes in user labels required to trigger the training of a new model, respectively. To ensure the user's model updated in reaction to their inputs but without causing wasted computation, the researchers set both thresholds to 4 based on empirical testing. Label Sleuth trains a SVM model for the user, allowing for fast retraining when new labels arrive. However, retraining still takes approximately a minute. After the training of a new model, the experimental platform determines which predictions by the new model are different from before, records them, and signals them to the user.


The predictions could change several times before the user finished assessing the tweets, they decided that each feed would consist of 26 tweets. With 26 tweets in each feed, each of the models Hidden and Visible would have seen 52 examples by the end of their life cycle. Additionally, providing free-text reasons for 3 tweets in each feed served the purpose of delaying users a little longer, allowing for the processing of model predictions.


Participants


The researchers recruited participants from Amazon Mechanical Turk, requiring them to have more than 500 HITs approved, an approval rate of over 98%, and the United States as their location. Additional eligibility criteria included being 18 years or older, occasionally reading news online, fluency in English, and being a US citizen or permanent resident. A total of 65 non-spammer workers participated. The researchers identified low-quality responses by examining participants' free-text responses to reasoning questions and the post-study survey. Four participants were paid but their submitted texts were unintelligible due to language errors. These cases were removed from the dataset as the participants did not fully understand the task.


Analysis Procedure

  • Due to a bug in the open-source AI system, 11 participants' AI models (Hidden, Visible, or both) failed to retrain after initial iterations. Researchers removed these participants' data from their analyses comparing assessments across conditions, resulting in data from 50 users for these analyses. However, for analyzing post-study survey responses, they retained data from 54 users, excluding only those whose Visible model experienced training problems, regardless of Hidden model failures.
  • For statistical tests predicting binary dependent variables, they fitted generalized linear models using the "glmer" function from the R package "lme4," with the family function "Binomial" and link "logit."
  • For continuous outcomes, used the "lmer" function to fit linear models.
  • In all models, included tweet and participant identifiers as random effects to account for variation attributed to unobserved characteristics of a particular tweet or participant.
  • In regression estimate tables, presented exponentiated coefficients (odds ratios) as measures of effect size for binomial dependent variables (logistic regression), and partial χ2 as a measure of effect size for linear regressions predicting continuous outcomes.
  • Additionally, they presented the marginal χ2 (variance in the outcome explained by fixed effects) and the conditional χ2 (variance explained by the entire model including both fixed and random effects) in all analyses.


Results


The results indicate that users are more likely to agree with the AI's assessment of a tweet when they are more confident in their own assessment. This relationship is illustrated in Figure 2.Therefore, as some participants hypothesized, "mistakes" made by a user's personalized AI could be attributed to the user's lack of confidence in assessing the tweets.



Figure 2 Jahanbakhsh et al.- The predicted values of user-AI agreement by the user’s confidence in their assessment. The higher the confidence of a user in their assessment, the more likely their AI is to correctly predict their accuracy rating. (Figure 6 in original paper)

 
Researchers analyzed participants' free-text responses through open coding to understand their views on the pros and cons of incorporating such an AI on social media platforms. The table summarizes these views, highlighting factors influencing the AI's perceived performance and participants' arguments for and against adopting a personalized AI for identifying misinformation on social media. Some participants expressed concerns about the specifics of the AI's implementation or deployment, particularly regarding the lack of trust in the AI's predictions and the absence of reasoning provided for its predictions. Others were worried about the potential manipulation of such a tool by social media platforms if it were operated by them. However, many participants saw potential benefits in using AI to identify misinformation, such as serving as a first-pass filter to filter out blatantly false content, providing a guideline for content inspection, or presenting alternative opinions in a structured form. 


When examining participants' responses, the researchers found that users tended to agree more with the AI's predictions when they were displayed to them compared to when the predictions were withheld. This suggests that users are influenced by seeing the AI's predictions when assessing content. Furthermore, they observed that the impact of seeing the AI's predictions on users' accuracy ratings increased over time, indicating that users became more reliant on the AI's predictions as they interacted with it more. The analysis also revealed that the interaction effect between the model and iteration was significant. Over time, the difference in user-model agreement between models Hidden and Visible grew larger, suggesting that users became more reliant on the AI when they saw its predictions.




Table 2 Jahanbakhsh et al.- Participants’ views in favor of and against adopting a personalized AI for identifying misinformation on social media. (Table 4 in original paper)



Figure 3 Jahanbakhsh et al. - The effect of the interaction between model and iteration on the predicted user-model agreement. Over time, the difference is user-model agreement across the two models increases (Figure 9 in original paper).


Figure 3 shows the effect of the interaction between model and iteration on the predicted user-model agreement. It demonstrates that over time (i.e., as the iteration increases), the difference in user-model agreement between models Hidden and Visible grows larger. This observation suggests that when users see AI predictions, they become more reliant on the AI over time.


Benefits and Challenges


The use of tailored AI for misinformation detection offers several benefits:

  1. Real-time Detection: It can quickly identify and flag misinformation as it appears on social media, helping to prevent its spread.
  2. Customized Solutions: By tailoring detection algorithms to individual users, it can provide more effective and targeted solutions.
  3. Reduced Human Effort: Automated detection of misinformation reduces the burden on human fact-checkers, allowing them to focus on more challenging cases.

However, there are also challenges associated with the use of AI for misinformation detection:

  1. Privacy Concerns: Tailored AI requires access to user data, raising concerns about privacy and data security.
  2. Algorithmic Bias: There is a risk that AI algorithms may inadvertently reinforce existing biases, leading to unfair or inaccurate assessments of content.
  3. Adversarial Actors: Those spreading misinformation are likely to adapt their tactics to evade detection by AI algorithms, posing an ongoing challenge.


The Future of Misinformation Detection


While tailored AI holds promise for preventing misinformation on social media, it is not a remedy. To be effective, it must be part of a broader strategy that includes human fact-checkers, community reporting, and efforts to improve media literacy. The future research and development would improve the accuracy and fairness of customized AI algorithms. By addressing these challenges, it can harness the power of AI to create a safer and more reliable online information environment.


Conclusion


The researchers examined the potential of incorporating personalized AI for determining content accuracy on social media. The aim was to empower users to make more informed decisions about the content they consume and share, thus addressing concerns about misinformation. The personalized AI learns from a user's assessments of content and predicts how the user is likely to assess other content. Through a user study, participants interacted with a personalized AI that learned their assessment patterns, predicted the accuracy of tweets, and evolved based on user feedback. The study also investigated whether users would be influenced by seeing the AI's predictions and found that they were indeed influenced, becoming more reliant on the AI over time. However, users' reliance on the AI was reduced when they provided reasoning for their assessments. This research paper highlights the potential of personalized AI in content moderation but also identifies challenges and areas for future improvement, particularly in terms of user autonomy and mitigating the influence of AI predictions on user judgments.


References


Farnaz Jahanbakhsh, Yannis Katsis, Dakuo Wang,  Lucian Popa, Michael Muller "Exploring the Use of Personalized AI for Identifying Misinformation on Social Media" in CHI '23: Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems.



-Nithiya


Viewing all articles
Browse latest Browse all 738

Trending Articles