Asking
Only logged in members can reply and interact with the post.
Join SimilarWorlds for FREE »

Anybody else notice that things really changed after the Covid Pandemic?

I don't know but to me it seems that the pandemic was the point of no return for the world, people seeming to have lost their minds, the world and America as a whole being too soft on crime which allowed it to Skyrocket and just straight up weird shit going on. don't get me wrong the world was messed up before but at least there was some version of Order. and the biggest kick in the pants all this INFLATION!!!!! which wouldn't be so bad if companies stopped being greedy and raise our wages.
This page is a permanent link to the reply below and its nested replies. See all post replies »
zonavar68 · 56-60, M
There is still a lingering standoff between the pro Covid Vax and anti Covid Vax crowds but hardly anyone asks and questioning a person's Covid status is pretty much forgotten.

In dating I think there's still a fairly strong undercurrent of women who straight out refuse to date men if the won't disclose their Covid Vax position or openly voice against it.

Masks linger on but few if any people still have them.

The WFH thing never affected my job.

Covid was a financial goldmine for the corporate/elite sector though.
ViciDraco · 41-45, M
@zonavar68 I keep my mask in my car. If I'm feeling ill, I'll wear it in public to reduce spread chances. It was a common thing to do in Asian cultures even before covid and I think it's generally a good and considerate idea. But yeah, most people have moved past them.