Recently, Google CEO Sundar Pichai has put a blog post detailing the ace search engine’s principles around Artificial Intelligence (AI). From the post it is quite clear that Google won’t be using AI technology to build weapons or mass surveillance systems.

According to Pichai’s blog post, “AI is computer programming that learns and adapts,” and it has tremendous potential to improve the lives of people. At the same time, he also admits that AI will not be able to solve all problems and its use will raise “equally powerful questions about its use.” The Google CEO clarifies that since the company is leader in the field of AI, it has a “deep responsibility to get this right.” In all, Google will have seven principles to guide their AI work and research. These principles will not just be concepts, but according to Pichai, “are concrete standards, which will govern our research and product development and will impact our business decisions.”

Elaborating Google’s AI principles further,  the company’s AI principles state that the AI will be socially beneficial, it will avoid creating or reinforcing unfair bias, will be tested for safety and will be accountable to people. AI from Google will also be tested for safety, have privacy principles built within the design, and uphold “high standards of scientific excellence.” Finally, the seventh principle notes that AI will be made available only for uses that are in accord with other six principles of the company. Along with this, Google will also evaluate as to when to make these new technologies available on a non-commercial basis.

Moreover, the blog post also clarifies that Google will not build AI technologies which will cause overall harm, weapons, including those which will “cause or directly facilitate injury to people.” It will also not allow AI technology from the company to be used for surveillance, which violate “internationally accepted norms.” Google will not allow its AI to be used against principles of international law and human rights.  The blog post also states that there is space for more voices in this conversation on AI principles, and Google will “work with a range of stakeholders to promote thoughtful leadership in this area, drawing on scientifically rigorous and multidisciplinary approaches.” Adding a touch of matured moral ethics to its AI principles, Google states that its AI will “respect cultural, social, and legal norms” across the countries where it operates.