“The Americans combine the notions of Christianity and of liberty so intimately in their minds, that it is impossible to make them conceive the one without the other.” Alexis de Tocqueville
In an earlier column, I wrote about the tendency of textbook authors to deny or denigrate the role of religion in their depictions of the founding of the United States. Historians like Professor Eric Foner teach their students that the Founding Fathers were able to embrace progressive ideas like freedom and equality because they viewed Christianity and the Bible as “outdated superstitions that should be abandoned in the modern age.”1
The truth is very different.
College history professors, like other left wing extremists, are loath to acknowledge that religion has played a positive role in the development of this nation; yet any honest portrayal of American history would have to acknowledge it. The rights and freedoms enshrined in the Declaration of Independence and the Constitution were, the Founders thought, quite literally sacred; having been bestowed on the human race by God Himself.
The American people of the late eighteenth century were more generally devout in their Christianity than the citizens of any other nation, and there is a reason for that. In America religion was not imposed on the people by government, it was freely chosen.