Associations to the word «Nigeria»

Wiktionary

NIGERIA, proper noun. Country in West Africa, south of the country of Niger. Official name: Federal Republic of Nigeria.

Dictionary definition

NIGERIA, noun. A republic in West Africa on the Gulf of Guinea; gained independence from Britain in 1960; most populous African country.

Wise words

Life has no meaning unless one lives it with a will, at least to the limit of one's will. Virtue, good, evil are nothing but words, unless one takes them apart in order to build something with them; they do not win their true meaning until one knows how to apply them.
Paul Gauguin