Skip to main content

Featured Story

Transformative Shift: COTI Leads the Future of Privacy with Ethereum Layer 2 Adoption

million, to fuel the holistic development of its ecosystem. This initiative is designed to support projects and developers who are aligned with COTI's mission of enhancing privacy, security, and scalability within the Ethereum ecosystem. The Ecosystem Growth Fund represents a significant investment in the future of blockchain technology and underscores COTI's dedication to fostering innovation and growth within the industry. Advancing Privacy with Garbling Circuits Technology COTI's transition to Ethereum Layer 2 signifies a strategic shift towards scalable privacy solutions within the blockchain space. The adoption of Garbling Circuits technology enhances the speed, efficiency, and security of COTI V2, positioning the firm as a pioneer in privacy-focused initiatives. Garbling Circuits technology opens doors to a wide range of applications, including privacy-preserving wallets, decentralized exchanges (DEXs), private AI training, governance mechanisms, and more. Part

Meta's Responsible AI Restructuring: A Strategic Shift in AI Development

Meta's restructuring of its Responsible AI team reflects a strategic shift in its approach to artificial intelligence development. In a move reported by The Information, Meta is taking a distributed approach in light of the increasing scrutiny around the role of AI in society. The company stated that it is integrating its Responsible AI team members into separate divisions across the organization, aiming to embed AI safety considerations directly into the development of core products and technologies.

The impact of generative AI has raised concerns worldwide, ranging from privacy issues to fears of an AI apocalypse. In response, Meta assures that its decision to disperse its Responsible AI team does not imply a lack of commitment to responsible AI development. The company emphasizes that it will continue to prioritize and invest in safe and responsible AI development.

However, critics like David Evan Harris, a former researcher at Meta, have raised alarms about the potential misuse of AI technologies. Harris, who previously worked on the civic integrity team at Facebook and managed research teams focused on responsible AI, has highlighted the potential harms of AI and sought ways to make it safer and fairer for society. He cited examples of how AI systems could facilitate housing discrimination, make racist associations, and exclude women from seeing job listings visible to men. Harris also expressed concerns about deepfakes and the possibility of open-source language models being used by malicious actors to spread misinformation. The now dismantled Responsible AI team played a crucial role in preventing these issues from occurring.

This restructuring coincides with Meta's initiative to streamline operations and align its AI development practices with the evolving landscape of societal concerns. By embedding AI safety considerations across different divisions, the company aims to ensure responsible AI development is integrated into the core of its products and technologies. As AI continues to play an increasingly prominent role in our lives, it is imperative for companies like Meta to address the ethical and societal implications of AI and work towards building AI systems that are safe, fair, and transparent.


Trending Stories