Doctors Aren’t Legally Required to Cure You – The Shocking Truth