June 28, 2012 --
Today the Supreme Court upheld the 2010 health care law in a dramatic victory for President Barack Obama. The lead up to today's decision has prompted debate between opponents and supporters of the Patient Protection and Affordable Care Act two years ago. Take a look at how we got to the health care system we have in place today.
Do Mexicans Have Better Health Care Than You?
Prior to the 20th century, nothing even close to what could be called a health care system existed in the United States. Although the Civil War had led to some medical breakthroughs in terms of surgical techniques and pain management, medical knowledge, techniques and treatment availability at the time left little hope that patients would actually recover from severe ailments. As NPR's Alex Blumberg and Adam Davidson point out, medical treatments may have been downright medieval at the time, consisting of potions. But at least it was cheap. "In 1900, the average American spent $5 a year on health care ($100 in today's money)," they note in their report.
How the Civil War Changed Modern Medicine
In 1912, Theodore Roosevelt was the first presidential candidate to get behind the idea of a national health insurance plan. Roosevelt ultimately didn't win election that year. Proponents of government-provided health care tried to press the issue through state initiatives, only to see their efforts fail in 16 states. Roosevelt's plan may have certainly been ahead of its time, particularly since there weren't that many services that doctors could actually provide patients during that era.
At the same time, however, developments within the medical community changed the face of the industry. The horrors of World War I led to advances in the areas of wound care, sanitation, pain management and more, according to an article published in the Journal of the Royal Society of Medicine. Hospitals in the United States began to widely adopt the practice of using antiseptics to sanitize their facilities, preventing the possibility of medical personnel or patients becoming exposed to infection. That decade also saw the introduction of the first employer group insurance contracts (though not specifically for health insurance) as well as the first physician service and industrial health plans.
In 1928, Alexander Fleming made one of the most important discoveries in the history of medicine: penicillin, a life-saving drug used to treat countless millions. It would be decades, however, before penicillin would be mass-produced. Fleming's discovery was the signature achievement in an era that saw medical treatment become more effective, and, as a result, expensive. The Great Depression also fueled concerns about affordability of medical treatment as millions of Americans suddenly found themselves out of work. In 1929, Baylor Hospital provided the first group health insurance plan in the United States through an agreement with Dallas-area teachers. The plan was the forerunner of Blue Cross. The effort wasn't just meant to be in the best interests of patients, but also the hospitals. Patient facilities saw more empty beds as fewer patients during the Great Depression could afford treatment without participating in these collective prepaid health insurance plans.
Ancient Nubians Drank Antibiotic-Laced Beer
As part of his push to create a social safety net for Americans during the Great Depression, President Franklin D. Roosevelt advocated the passage of national health insurance. Roosevelt pushed ahead with efforts to pass Social Security first, a bill which intentionally omitted any mention of medical care to ensure its passage. Harry Truman attempted to carry on Roosevelt's legacy in 1945 by calling on Congress to create such a program. His efforts failed, partly due to criticism by the American Medical Association (AMA), who called the plan "socialized medicine." In this photo taken in 1937, First Lady Eleanor Roosevelt examines a chart of enrollment of health care insurance plans.
Like its predecessor, World War II would lead to new medical advancements, including the widespread adoption of antibiotics and the use of ultrasound. The war would also have a similar effect in terms of the spread of employer-sponsored health plans. Because the nation was in a state of emergency and had a legally mandated wage freeze as a result, employers had to attract workers to assist the war effort by providing them with benefits, including health insurance. Tax laws passed between 1943 and 1945 also gave breaks to employers who provided insurance to their employees, which gave businesses all the more incentive to offer coverage. Following the war, employer-sponsored health insurance became common. In 1951, around 77 million Americans had some kind of coverage, according to an insurance industry trade group. That era also saw one of the most celebrated medical achievements in history: Jonas Salk's polio vaccine.
NEWS: World War II
Although health insurance was widely available to employed Americans in the mid-20th century, the unemployed and the elderly were often excluded from these plans. President John F. Kennedy campaigned on the issue of insuring these groups. President Lyndon B. Johnson succeeded where Kennedy left off, securing the passage of a bill through Congress creating Medicare and Medicaid. At the bill-signing ceremony, shown here, Johnson presented former president Truman with the nation's first Medicare card. Within the medical industry itself, an increasing number of doctors began specializing in certain fields of medicine rather than acting as general physicians. By 1960, more than two-thirds of doctors reported themselves as full-time specialists, rather than general practitioners.
Starting with Richard Nixon in 1970, presidents have offered successive plans for covering the nation's uninsured, but they have have stalled for different reasons. In 1974, Nixon put forward a plan to cover all Americans through private insurance, only to have the Watergate scandal force him out of office. An economic crisis prevented Jimmy Carter from pushing forward with a national health plan. Congress late in Reagan's second term attempted to expand Medicare, only to have the law repealed the following year. Bill Clinton had a 1,300-page health care reform bill that was never even taken up for a vote in Congress. Since Nixon's presidency, health care costs have continued to rise, often outpacing inflation. This increase is due to a number of factors, including the increased use of new medical technologies for diagnosis and treatment. The Patient Protection and Affordable Care Act signed by President Barack Obama was intended to cover the 30 million Americans who live without health insurance, according to the bill's authors. It has been the most far-reaching piece of health care legislation since Johnson's signed the legislation creating the Medicare and Medicaid health care programs.
NEWS:Think You'll Live Long? Think Again
Regular exercise may lower the risk that a person's sense of smell will fade, which tends to happen as people age, a new study finds.
The study included about 1,600 people ages 53 to 97 who didn't have any problems with their sense of smell at the study's start. They were followed for up to 10 years. Researchers tested the participants' ability to detect eight odors, including chocolate and coffee, three times over the course of the study. Participants also reported their exercising habits.
During the study period, about 28 percent of people developed impairment in their sense of smell, and the results showed a link between exercising habits and smelling ability, after adjusting for age and gender.
"Participants who reported exercising at least once a week long enough to work up a sweat had a decreased risk of olfactory impairment," the researchers wrote in the study, published today (Oct. 17) in the journal JAMA Otolaryngology-Head & Neck Surgery.
The results also showed that the more people exercised, the lower their risk was of developing olfactory impairment.
The sense of smell serves a number of purposes: It helps people identify toxic chemicals, and enjoy the flavor of foods. The sense of smell, or olfactory function, may decline as people age, and men are more likely than women to see a decline in their smelling abilities.
Studies have found that about 25 percent of people ages 53 and older have an impaired sense of smell. Risk factors for developing this decline include smoking and having nasal polyps or a deviated septum.
It's not clear how exercising may protect a person's smelling ability, but one possible explanation could be that exercise improves either brain function or overall general health, the researchers said.
The study found an association between exercise and maintaining a better sense of smell, but it doesn't prove a cause-and-effect relationship. It is also possible that people who exercise are generally healthier, and their lifestyle and other behaviors have a protective effect on their olfactory system, the researchers said.
More From LiveScience:
Minty? Lemony? Humans Can Smell 10 Types of Odors
Tip of the Tongue: The 7 (Other) Flavors Humans May Taste
7 Common Exercise Errors And How to Fix Them
Copyright 2013 LiveScience, a TechMediaNetwork company. All rights reserved. This material may not be published, broadcast, rewritten or redistributed.