China’s regulatory requirements shine a bright light on one of the most interesting issues facing AI regulators: the alignment problem, which refers to the challenge of ensuring that AI systems (especially those that operate autonomously) act in a manner that aligns with human values, intentions, and objectives. Which raises the question: Whose human values?
A core aspect of the alignment problem is that it’s not always straightforward to specify what we want an AI system to do, and it’s even harder to specify what we want in a way that leaves no room for misinterpretation. If a machine learning system is given a goal without the proper context or constraints, it may find solutions that technically meet the goal but violate the spirit of what was intended.