President Trump recently caused a furore by allegedly referring to Haiti and African countries as ‘shitholes.’
But despite media condemnations/ obfuscations in both Left and Right wing media circles, is he wrong?
I can hardly imagine these critics would voluntarily trade in their cosy Western life for Haiti. And for all the celebrities who swore they would leave the United States under a Trump Presidency, which one of them picked Haiti?
Which begs the question: why is Haiti such a desolate, impoverished place to live? Could it be due to Western imperalism and oppression?
Considering that Haiti was once considered to be the ‘Paris of the Antilles,’ only for it to fall into disrepair in the centuries after its remaining whites were exterminated, this answer is insufficient.
So what else could possibly be blamed?