Meta's Shift Raises Concerns About AI Responsibility
Meta, the company behind social platforms like Instagram and WhatsApp, is making changes that have raised eyebrows. They've moved their Responsible AI team, which focuses on ethical AI use, to join the generative AI team. This comes amid issues like mislabeling users and generating inappropriate content.
In the past, Meta faced challenges, including layoffs and the discontinuation of a fact-checking project. David Harris, a Metaverse project lead, expressed worry about the company's ability to address these issues. The Responsible AI team, established in 2019, previously faced difficulties and had limited influence.
The decision to disband the Responsible AI team is a tricky move. On one hand, having a separate team for responsible AI can be beneficial, but it might lead to delayed action. If they were involved earlier, they could catch problems before they escalate. An example is Meta's model Galactica, similar to ChatGPT, which had issues distinguishing truth from fiction.
Despite layoffs and controversies, Meta is focusing on generative AI, aiming for efficiency and cost-cutting. Mark Zuckerberg declared 2023 as the "Year of Efficiency," but recent events, including team restructuring, have sparked concerns about Meta's commitment to safety and ethics.
The changes highlight how companies prioritize meeting Wall Street's demands for efficiency, sometimes at the expense of safety and ethical considerations. Former employees and observers are questioning Meta's approach in the face of these shifts.