Usually when I hear mention of European men, I think of fags wearing tight jeans. Not Iron Age nordic or medieval crusader types.
Is it true that in the United States American men are forced to work for niggers and receive pay in the form of cheeseburgers?
Do you see any niggers that own huge American based companies?
No because the kikes own the companies they just put niggers into management.
Also obama.
technically the richest man in the world is African American.
(post is archived)