Absolutely. Before the blm stuff in 2020, and the woke stuff in general that took over the country around 2016, it was quite common to see confederate flags here. And i dont even like in a rural area, i live in a liberal city.
But iirc after 2020 the confederate flag is banned
people talk a lot about slavery and i understand why but literally the only reason people flew that flag was as a way of saying "i love the south" because there is no other symbol that represents the american south as a whole. And because it looks cool AF
Anywho I'm on a whole side tangent now. Yes we absolutely consider ourselves southern. Country music, cornbread, NASCAR, saying "y'all", common courtesy and general kindness, everything. But southern culture is on its way out here because there are so many northerners that move down here and refuse to assimilate. For some reason, so many of them move down here and are complete assholes to everybody and they constantly talk about how much better it was up north. So if its so much better, then why do they deliberately move down here? Even my dad is a northerner, but he at least somewhat respects the culture though (to an extent).
southern culture is also disappearing because theyre importing so many darkies here. In my short life (20 years old) ive witnessed white people become a minority in my hometown
its sad to watch man, theres no more trust or kindness here now.
Anywho thats way more than you cared to hear, and this comment reads like ass because im so tired lmao, but TLDR yes we are absolutely a southern state