How do we as humans make sure AI doesn’t turn evil and take over the world?

Jennifer Turliuk
2 min readApr 11, 2023


The notion of AI turning evil and taking over the world is often portrayed in popular culture, and many think this scenario is likely. As AI continues to evolve and become more integrated into our lives, there are legitimate concerns about the ethical and social implications of these technologies. Here are some ways in which we can ensure that AI is developed and used in a responsible and ethical manner:

  1. Ethical Guidelines and Standards

Developing ethical guidelines and standards for the development and use of AI will be essential to ensure that these technologies are aligned with human values and serve the greater good. Governments, industry leaders, and academic institutions should work together to establish clear and comprehensive ethical standards that reflect the diverse needs and perspectives of society.

2. Transparency and Accountability

Transparency and accountability are crucial for ensuring that AI systems are fair, unbiased, and free from harmful consequences. Developers should be transparent about how AI systems make decisions and be accountable for any negative outcomes that may result from their use.

3. Diversity and Inclusion

Ensuring diversity and inclusion in AI development teams is essential to avoiding biases and ensuring that AI systems work for all individuals and communities. Diversity and inclusion will help to ensure that AI reflects the diverse perspectives and needs of society, and that its benefits are widely shared.

4. Regulation and Oversight

Regulation and oversight of AI development and use will be important to ensure that these technologies are used in a responsible and ethical manner. Governments should work to establish clear regulatory frameworks that provide oversight of AI systems and protect individuals from harm.

5. Education and Awareness

Educating the public about AI and its potential implications will be important to promote responsible use and mitigate fears and misconceptions about AI. Governments, academic institutions, and industry leaders should work together to educate the public about the benefits and risks of AI, and promote responsible use.

6. Collaboration and Partnership

Collaboration and partnership between industry, government, and academia will be essential to ensure that AI is developed and used in a responsible and ethical manner. By working together, these stakeholders can share knowledge, expertise, and resources to promote the responsible development and use of AI.

In conclusion, ensuring that AI is developed and used in a responsible and ethical manner will require a collaborative effort from industry, government, academia, and society at large. By establishing clear ethical standards, promoting transparency and accountability, ensuring diversity and inclusion, regulating AI development and use, educating the public, and promoting collaboration and partnership, we can ensure that AI serves the greater good and benefits society as a whole.

[Written with ChatGPT, so take it with a grain of salt]



Jennifer Turliuk

Investor | Entrepreneur | Technologist | Author | Speaker | Climate Tech, Edtech, Impact, AI. Partner, Climate Angels & Climate Capital Syndicate