Meta recently suffered consecutive defeats in lawsuits in New Mexico and Los Angeles, with juries determining that Meta was aware of the potential dangers of its products but failed to regulate them. These two verdicts have raised concerns among experts that the tech industry, in an effort to evade legal responsibilities, may further reduce the number of social researchers, suppress research on artificial intelligence models, and psychological safety assessments, thereby exacerbating the risks for consumers.
Former Facebook executive exposes internal research as strong evidence
Meta, formerly Facebook, has employed a large number of social science experts over the past decade to analyze the impact of social networks on users, showcasing the company’s willingness to take on innovation risks. However, recent rulings indicate that these studies, originally intended for product improvement or public relations, are now being cited in court as evidence of “willful misconduct.” Former Meta executive Brian Boland pointed out that the findings of internal research often contradict the image the company projects externally; internal investigations showing that teenage users on Instagram are susceptible to sexual harassment have become favorable evidence for plaintiff lawyers accusing negligence. These studies, once regarded as part of corporate social responsibility, have now become a heavy burden for business owners in legal battles.
Experts worry that removing researchers will weaken assessments
Since former product manager Frances Haugen revealed a large number of internal documents in 2021, Meta has significantly tightened control over internal research. The leaked documents confirmed that Meta was already aware of the potential negative impacts of its products, marking a turning point in global regulation. The non-profit organization Children and Screens noted that many tech companies are adjusting their strategies to mitigate legal risks, eliminating research that could be detrimental to the company. According to reports, Meta and other tech giants are gradually reducing the size of their research teams and even removing data tools available for third-party researchers. Experts are concerned that if companies continue to view safety research as a burden, it will weaken their ability to conduct fair assessments.
Insufficient AI safety research poses threats to mental health
Companies like OpenAI, Google, and Meta have invested substantial resources in researching models, but there is a noticeable gap in studies addressing their psychological impacts on consumers. Kate Blocker analyzed that current AI research and development primarily focuses on technology rather than the long-term effects on the psychological development of teenagers and children from chatbots and digital assistants. Experts worry that AI is repeating the mistakes of social media; by removing researchers due to concerns that research findings may become unfavorable evidence in court, there is a risk of the public being insufficiently aware of the potential harms behind AI products, leading to physical and mental injuries.
This article “Meta’s defeat could accelerate the tech industry’s reduction of social psychology research, increasing consumer risks” first appeared in Chain News ABMedia.