I believe that America is much more politically correct than it was when I was born. But does this mean that the american public is necessarily less racist or sexist? Not at all. In the 21st century, people are extremely conscious of what they can say about a person of the certain race or gender. The unwritten rules of what you can say and do are essentially public knowledge to everyone. You immediately know if someone has overstepped the invisible boundary when they say something that is offensive to another race.
However, this does not at all mean that people’s feelings have changed. Just because the country is more politically correct doesn’t mean that the racist or sexist feelings of many Americans have faded or subsided in our era. I have no doubt in my mind that people still have negative feelings toward people of different races.
The only difference nowadays is that they are not allowed to openly share these feelings. While the country has obviously progressed significantly in terms of racial acceptance, there will obviously always be people who dislike other races. Barack Obama becoming the first black president was big, but then again the ill feelings towards races have just shifted. People blame hispanic people for taking American jobs, and many people feel uncomfortable sitting next to someone who appears middle-eastern on a plane. America is more politically correct, but no more accepting than it was when I was born.