What do you think?
Politics used to be about running the country, now it's about acting like a buffoon and putting on a show while the corporations eat out our substance.
Doctoring used to be about curing diseases, now they are convincing people they can be healthy at any size.
The democrats used to be about the worker and environmental protection. Now they are about Sodomy, anal rape, penis torture and the right to kill children indiscriminately like bugs.
Republicans used to be about states rights and compartmentalized government: now they are about opening the ass for Israel, dismantling the governments restrictions on fraudulent corporations and staging violent revolutions against the state.
School used to be about teaching you to write in cursive and times tables, now they do neither and teach you how to be a whore and a fag.
Since everything is going to shit and I'm not even 30 years old yet, what colleges would you say are still worth attending if any?
What colleges are not:
1: promoting institutionalized sodomy (LGBTQIAPF) through their official administration
2: Suppressing views they don't like, especially ones like "maybe we shouldn't kill our young."
3: Won't call the cops if you protest against Israel (funny how the NYPD sits around eating donuts while people get shot to death in the streets, but once an israel protest comes up OH BOY SCRAMBLE THE FLEET!!)
4: Nigger worship
5: Chink infestation. They colonize the place like bugs and from what I've heard they don't like whitey too much and then the school starts catering to their needs above the needs of whitey.
I feel like Bill Maher and all his Jew friends hyped the whole "Harvard is woke" when the protests happened to discredit the protesters. But is it true?