The problem with existing or mainstream religions is that they are life-denying and don't recognize realities such as race, and this is why they are harmful. There seems to be a general trend with any sort of organization, party and religion where, if they are not explicitly pro-White, they will fall prey to the logic of anti-White ideology. There are so many religions / philosophies out there which are founded on the following sorts of ideas:
>life is suffering, everything is impermanent (Buddhism)
>this world is just a test for a perfect afterlife (Abrahamics)
>I will be reborn endlessly in this world until I escape (Hinduism, Buddhism again)
>the world of the senses is an illusion, nothing changes, all is one (Eleatics)
Et cetera. All of these views downplay or even deny the reality we find ourselves in for reasons that I believe we have no reason to accept, even if parts of each may
have some partial truths within them. Like David Lane said, God and religion are separate things, often conflicting. Evidence for divine handiwork is all throughout the natural world. Religion is the creation of mortals, and can hence help or destroy a people depending on the circumstances, and reflects their outlooks and experiences in life. If one really wants to learn what God / Nature, look to Nature. This can scare the weak and subhuman, as the fiction of God as "all good" from the human perception is utterly false, that force or intelligence than men have called God in reality can be quite impassive, indifferent or cruel from the limited view of mankind.