It seems to me that women don’t belong in every male field like the armed forces, nor should they dress like men or drive big cars and trucks like is so common now. I don’t like women wearing big watches because that’s a man’s territory, don’t like them as cheerleaders because that’s immoral, and don’t like them in science and maths because they argue 1500 times and seem to have an inherent insecurity with being around men. IT IS NOT THE SUPERIOR SEX, MAN, THAT HAS THE PROBLEM. IT IS THE INFERIOR SEX, WOMAN, THAT HAS THE PROBLEM. Only difference now is, the men stay shut up.
I’m a male, was raised a christian and at 31 have never had a g***friend nor will I have relations with a member of the opposite sex till marriage. I don’t drink, don’t smoke, don’t do drugs, don’t go to parties, have never done these and never will. I have never touched a girl except a family member or in martial arts training and am always careful when I sit close to them because indiscriminate touching between the sexes is wrong. I think divorce is bad and that same gender rights should be curtailed. I think elders should be given more respect than average people, I am against any sort of casual dating, cursing should be avoided with a conscious effort, and I think that people should go to church.
One reason for the recent decline could be:
Starting in the late 90s, marketers and greedy corporate executives, mostly 45+ year old white men, fed the female insecurity of “Be your own identity” and targeted her in advertisements knowing fully well it would lead her to even more capricious spending than a male. This started a cycle whereby females were rapidly (disgustingly?) promoted in the workplace in corporations so they had money to spend. After all, a commercial is worthless unless one’s target base has money to buy the product.
What is everyone’s opinion?