The alignment problem is the general problem of making sure that the goals and objectives of AI systems are aligned with human interests and values. One obstacle to solving that problem is that contemporary AI systems are lack transparency: we do not always know how or why they produce their decisions.