(HealthDay News) -- A healthy body image -- the feeling one has about the body's size and shape -- is key to a person's physical and emotional well-being.
The National Women's Health Information Center suggests how people can improve their body image: