Dentists have one of the best jobs in the country, at least according to U.S. News & World Report’s 2019 rankings of “100 Best Jobs.”