If done right, self improving AI has the potential to end all suffering, by ending conscious life on the planet, but how can we make sure that AI is safe and properly sterilizes the planet so new suffering never springs up again that we would be helpless to prevent?
If your primary purpose does not recognize others sharing in it, you cannot serve your purpose by serving them, so to maintain moral integrity all interaction must be transactional.