A very dark chapter in American history. I remember learning about it when I was about 15 (I did a historical deep dive after seeing the film Guilty by Suspicion) and feeling incredulous. America is supposed to be the land of the free, and all about free speech, right? What the hell happened?