When thinking about safety risks from ML, there are two common approaches, which I'll call the Engineering approach and the Philosophy approach:

While Engineering and Philosophy agree on some things, for the most part, they make wildly different predictions about what the key safety risks from ML will be and how we should address them:

In my experience, people who strongly subscribe to the Engineering worldview tend to think of Philosophy as fundamentally confused and ungrounded, while those who strongly subscribe to Philosophy think of most Engineering work as misguided and orthogonal (at best) to the long-term safety of ML