When did medicine become political? Good question. When doctors became employees rather than independent practitioners. When medical schools went woke. When state medical boards threatened to pull licenses if doctors dissented from official dogma. That’s when.