Do you think nature works against us whites ? I mean blacks are clearly physically superior. Their genes overtake ours. Socially they are gaining terrain. Our women crave them even the ones who claim they would never touch a black man. We whites are becoming minorities in our own countries and most of us just jerk off to the idea of going extinct. Is there any future for us whites or will we become history ?