Associations to the word «Nigeria»

Wiktionary

NIGERIA, proper noun. Country in West Africa, south of the country of Niger. Official name: Federal Republic of Nigeria.

Dictionary definition

NIGERIA, noun. A republic in West Africa on the Gulf of Guinea; gained independence from Britain in 1960; most populous African country.

Wise words

The chief difference between words and deeds is that words are always intended for men for their approbation, but deeds can be done only for God.
Leo Tolstoy