Ask Question
27 July, 06:02

How did world war 1 change women's roles in the united states? a). Women received greater educational opportunities b). Women fought alongside men in the military c). Women replaced men in the workforce d). Women earned more money than men

+3
Answers (1)
  1. 27 July, 06:07
    0
    c

    Explanation:

    im smart
Know the Answer?
Not Sure About the Answer?
Find an answer to your question 👍 “How did world war 1 change women's roles in the united states? a). Women received greater educational opportunities b). Women fought ...” in 📗 History if the answers seem to be not correct or there’s no answer. Try a smart search to find answers to similar questions.
Search for Other Answers