Growing up, whenever me or my sister complained about having to get naked for the doctor, my parents would say that everyone has to get naked for the doctor, it's just what you do. It was made to feel like a normal part of life and childish to complain about. It seemed so opposed to everything we were told about nudity being something private and something to be ashamed about. It made getting naked for the doctor seem so much more arousing. Did anyone else have a similar experience growing up?