I know this has been a thing growing since the 1970's, but why oh why do the women want to push fathers and husbands away from the family and from children?
It is so sick!
Don't they know what they are doing?
Do they realize they cause great harm to children by destroying the family unit?
WHY DO THEY DO THIS? AND IT HAS BECOME SO POPULAR TO HATE MEN!
How did such misinformation become so prevalent?
I assume it has to do with power. It is odd and sick that women want to harm others in this way. All the while they pretend they are on the pedestal, with perfect moral virtue.