I just can't get over the gall non-whites have. These Western Countries are not theirs yet they think they have some kind of natural right to tell us that we have to accept them and we're bad people if we don't just let them come in and take over and rule over us and own us. It's really bizarre to say the least and I just can't understand why any White Person would look at this and think that it's the right thing to do to give up their Country, Culture and Civilization.