I understand that world wars have shaped perception and society in europe, but rolling over on your back and surrendering to a bunch of muslim invaders isn't in the best interests of europe.
Apparently, they have. Have Americans really become this neutered, effeminate, and down right suicidally stupid?
We're not that many years behind the Europeans.