bn:03601053n
Noun Concept
PT
No term available
EN
Women's colleges in the Southern United States refers to undergraduate, bachelor's degree–granting institutions, often liberal arts colleges, whose student populations consist exclusively or almost exclusively of women, located in the Southern United States. Wikipedia
Relations
Sources