Meaning of west germany
Definition of west germany
(noun)
a
republic
in
north
central
Europe on the North Sea;
established
in 1949 from the zones of Germany
occupied
by the British and French and Americans
after
the German defeat; reunified with East Germany in 1990
Other information on west germany
WIKIPEDIA results for
west germany
Amazon results for
west germany
Tweet