Did Germany lose WW1?
Outcome of World War I for Germany
Germany emerged as the defeated force in World War I. Following the conflict, the 1919 Treaty of Versailles was established, wherein the allied nations, including the United States, Great Britain, France, and others, levied harsh territorial, military, and economic penalties on Germany.