Meaning of dermatology

Definition of dermatology

(noun) the branch of medicine dealing with the skin and its diseases

Other information on dermatology

WIKIPEDIA results for dermatology
Amazon results for dermatology