We’ve all been told—especially through the COVID-19 pandemic—that hand washing is important, and new survey results show that more than nine in 10 Americans believe this. According to Bradley ...