As the Covid-19 pandemic continues to surge in the United States, Americans are becoming more aware of the deficits in their health care delivery system. Invisible to many, however, is the rapidly expanding role that private equity is playing in health care — especially for women.