Science, Gender, and the Balanced Life

People often ask how I “broke through the glass ceiling” to succeed in a field that is predominantly male and overcame the barriers women face in science. They don’t always like my answer, because it’s not about gender. It’s about learning how to be effective as a member of a minority in different contexts, understanding the peculiar structure of academic research, and accepting the choices and tradeoffs facing both women and men in balancing work and life in this field.

From my initial interest in science in high school, I never felt held back, overlooked, or underappreciated because I’m female. Quite the opposite: Perhaps because people in science often are so passionate about their work, they may be gender-blinded by the thrill of sharing their fascination and wish to encourage anyone who is interested. And I was.

My enthrallment with science began with thoughts of becoming a physician. That goal changed after a year off after college, when I worked in the Columbia University lab of the eminent neurobiologist Eric Kandel, who went on to share the 2000 Nobel Prize in Physiology or Medicine for his research on the physiological basis of memory storage in neurons.

Kandel and his colleagues were so encouraging that they inspired me to pursue a career in scientific research rather than medicine. That led to successful graduate and postdoctoral research work, including a Ph.D. at Yale in biology/ neuroscience and research positions at the Salk Institute and the University of California at San Diego.

My path was set. And I had a lot of support from both men and women in science—and owe them a lot.

But a funny thing happened along the way: Despite loving research in all its aspects, I became unsettled by the choices necessary to pursue a career in academic science. First and foremost, one must select a specific area of focus. That requires a measure of healthy monomania, a conviction that one’s chosen question and approach are more interesting, valuable, compelling, and likely to succeed than others’.

But as much as the scientific problems I was working on were enjoyable and inspiring, equally (and sometimes more) engaging and exciting were the projects and results of other scientists. The need to choose a single focus area, in fact, panicked me. This culminated in a pivotal, life-changing experience when, as a postdoc, I received a phone call with good news about funding of my grant application and realized I was more depressed than elated.

If a research scientist is not ecstatic about rare positive news about funding, it is most definitely a sign not to ignore.

After much soul-searching, I decided to leave research and began looking for career opportunities where my breadth of interest would be an asset. An ad for an editorial position at the scientific journal Neuron led me to publishing and eventually the position of editor-in-chief. The pursuit of an editorial career satisfied every need to remain deeply engaged with science and scientists while indulging in broader scientific issues. And although the demanding schedule of academic research was not an issue for me personally when considering this change in career direction, the structure and teamwork nature of editorial work do provide a framework of shared goals, expectations, and flexibility necessary to support a more healthy work/life balance.

My new path led to my leadership today of Cell and Cell Press. Again, all the while I was encouraged, or at least never discouraged, by male and female managers, mentors, and colleagues.

Science might be served better by a system that favors diversity, teamwork, balance, broad intellectualism, civic responsibility, and healthy work/life balance.

A minority of one

All that said, throughout my professional experiences, from the lab to Cell Press, I have often found myself in the minority; sometimes a minority of one. The only woman. The only junior scientist or editor in a room of distinguished leaders. The only American in a room of Europeans. The only academic in a room of businesspeople. More recently, the only business-minded person in a room of academics.

In all of these cases, effective discussion, successful collaboration, and recognition of the value of my contributions always required a good deal of listening and asking questions, putting myself in others’ shoes, and cogently explaining my perspective and ideas. I’ve found that closing a “gap,” gender or otherwise, requires mutual respect, shared goals, and a positive teamwork approach. Whether or not this attitude measurably contributed to my career success, it certainly made work more fun and rewarding.

If my experience is illustrative, then it’s fair to ask, why is there a gender gap in science, especially at the advanced levels? As the 2011 U.S. Commerce Department report Women in STEM: A Gender Gap to Innovation concludes, women remain vastly underrepresented in science, technology, engineering, and mathematics (STEM)–related jobs and among degree holders, and women with STEM degrees are less likely to work in STEM jobs, even though the gender wage gap is smaller than in non-STEM jobs. What’s going on here?

Let me suggest a perhaps provocative factor: that the gap is driven by the unusually individualized and competitive nature of academic research. Science trains far more students and postdocs than the number of faculty positions that are available. Growth in government research funding in many countries is limited. Thus, the field is sharply competitive for grant dollars, faculty positions, and stature, and it tends to reward individual over team achievement (no ambitious postdoc ever got a grant or job because she was a good team player or helped another scientist get a key publication). Moreover, no formalized structure tells a scientist what question to work on, what approach to take, or how many hours to invest. No principal investigator I know ever told a postdoc that he or she really should take a vacation.

Right or wrong, these factors are the reality of academic science today and of how scientists are rewarded and advance. Those who stand out and rise above tend to work long days, nights, and weekends, often traveling and presenting constantly. It can be hard for someone who wants a diversified life to compete against someone who’s willing to be in the lab 20 hours a day or on the road for months.

As a result, devotion to science research and pursuing the highest achievement makes seeking a work/life balance (e.g., family time) especially difficult, as is true in many competitive fields. The unique demands of science research exacerbate the challenge. The individualized nature makes it harder for parents and employers to distribute the workload when family duties call. What’s more, in my experience— forgive me—men still tend to be more comfortable with forgoing family time for work. Because the field is so competitive and fast-moving, women who take time for family can lose ground quickly.

Science demands hard choices

So to me, the gender gap is not necessarily a gender issue. That is not to declare the field completely free of gender or other biases, overt or covert, that should be addressed as they occur. Some of my colleagues note other challenges and solutions, including the need for women scientists to serve as role models to inspire young women to follow in their footsteps or mentor young professionals.

On a separate front, many institutions are striving to retain talent by helping both women and men balance work and life. Policymakers and advocates are calling for more generous family leave measures in the United States, similar to what other countries have. And many women and men I know do already achieve the balance they seek and happily thrive on both sides of the scale.

That’s my point: Scientists, like everyone else, understand that life is about making choices, accepting tradeoffs, and adjusting expectations. Plenty of research scientists I know who do not wish to shortchange family life (or other endeavors) for this demanding field are realistic; they understand that they also may not be on the fast track for a Nobel Prize or heading a major research institution. They accept that and are happy to adjust their goals in carving out a successful, productive research career with a comfortable sustainable position in the marathon, but not aiming to be the first to cross the finish line.

I hate to say it, and some might chafe at this notion, but a person who chooses work over other life pursuits may, by definition, do better in that field. That may not be the standard we wish to set. Science might be served better by a system that favors diversity, teamwork, balance, broad intellectualism, civic responsibility, and healthy work/life balance. If so, then collectively we need to embrace and pursue that goal with measures for both men and women that foster and reward those attributes in gaining funding, jobs, and status.

In the end, perhaps Women’s History Month is really about choices, for women, men, and the profession, about the system and life we want. It’s fair to assume that Marie Curie and the other 2013 honorees in the National Women’s History Project Women Inspiring Innovation through Imagination: Celebrating Women in Science, Technology, Engineering and Mathematics made their own tradeoffs. As Dr. Curie once said, “I have frequently been questioned, especially by women, of how I could reconcile family life with a scientific career. Well, it has not been easy.”

Schools Alone Cannot Close Achievement Gap

Gaps in student achievement are well documented. Members of some ethnic minority groups and low-income students consistently perform less well on achievement tests than their peers do. For example, a more than 20-point gap between white and Hispanic students G on National Assessment of Educational Progress tests in reading and mathematics has not changed significantly since 1990, and the gaps between black and white students have followed a similar pattern. Recent work has highlighted increases in the gaps among children of different income levels. Achievement gaps show up before children enter kindergarten: Children in the highest socioeconomic group entering kindergarten have cognitive scores 60% higher than those of children in the lowest socioeconomic group. The gaps in test scores and other measures persist throughout K-12 education, and corresponding gaps in high-school graduation rates, college matriculation and completion, and lifetime earnings demonstrate the impact that poor academic achievement has on young people’s lives. But, as Christopher Jencks and Meredith Phillips noted more than a decade ago in a study of test scores, such gaps are not “an inevitable fact of nature.”

A large volume of research has documented the associations between educational outcomes and factors associated with income and family and cultural background—most recently the collection Whither Opportunity?, edited by Greg Duncan and Richard Murnane and For Each and Every Child, a report from the Equity and Excellence Commission of the U.S. Department of Education. A picture is beginning to emerge of the specific ways in which economic resources influence education, but it has not yet resulted in policies that significantly narrow the gaps.

Schools clearly make a big difference. Research has established that the students most likely to lag behind academically are those who attend schools with less-qualified teachers and poorer resources. The rigor of the curriculum as it is implemented, the quality of teachers, class size, and teacher absence and turnover all have been shown to influence outcomes for students. In other words, what happens once children enter school may support those with disadvantages, or may perpetuate or exacerbate the gaps. (These issues are discussed in detail in a companion article by Natalie Nielsen.)

But there are other factors struggling students frequently share. For example, students whose families are not stable and supportive (those who change schools frequently, whose parents do not participate actively in their education, or whose families are disrupted by substance use or crime) are more likely to struggle in school. So too are students who live in poverty; whose neighborhoods are stressed by unemployment; and who feel unsafe at, and on the way to and from, their schools. The lack of adequate health care and adequate nutrition and untreated medical and mental health problems also are associated with school problems. Each of these sources of disadvantage may significantly impede a child’s academic progress, and these risk factors tend to cluster together, exacerbating their effects.

These important influences on children’s development have been the subject of considerable research, but less progress has been made in directly linking disadvantage that originates outside of school to educational outcomes. The National Research Council (NRC) has produced reports that synthesize research in areas that have important implications for the educational progress of children and adolescents. This article summarizes some of the messages from a body of work produced by the NRC since 2000 that offer insights about the out-of-school factors that may influence educational outcomes.

Produced primarily under the aegis of the Board on Children, Youth, and Families and the Committee on Law and Justice, these reports were not focused on addressing education issues and for the most part do not identify causal connections between out-of-school factors and academic achievement. They do, however, identify mechanisms by which specific sources of disadvantage impede students’ lives at school. As a group, they clearly point to important consequences of out-of-school factors for educational progress and to intersections among the different specific sources of disadvantage. The messages in these reports fall into a few broad categories.

Health and development

Research has vastly expanded understanding of development, particularly in early childhood and adolescence, and points to significant implications for education. Perhaps most important is the recognition that healthy development is a complex and interactive process that encompasses cognitive, psychological and emotional, biological, and behavioral processes, as well as environmental influences.

For example, as experts in early childhood care and education know, children who are ready for kindergarten have had the opportunity to develop the skills on which literacy and mathematical thinking will be built, but also the opportunity to form secure attachments and to develop social and self-regulation skills, so that they can develop successful relationships with their teachers and peers and take advantage of the classroom environment. The kinds of factors that frequently interfere with this development are closely associated with poverty: insecure housing and nutrition, unstable parenting and disruptions in relationships and care, and inadequate medical care.

The presence of extreme disadvantage, particularly when there are multiple sources, may be most harmful in the first years of life. Early experiences shape brain development, researchers have found, and the brain will adapt itself either to positive or to negative experiences. This means that infants and very young children who experience highly stressful family situations and associated risk factors can be permanently affected. These children will enter school at a significant disadvantage as compared with their peers and will present very different challenges to their teachers.

The early years are critical, but brain development continues into early adulthood, and it continues to interact with other factors. Researchers who study aspects of adolescence from different disciplinary perspectives are increasingly recognizing the importance during that phase of life of interactions among brain development, other biological processes, and social and environmental influences. Just as infants’ brains are particularly vulnerable to clusters of risk factors, older children and adolescents may respond to the same challenges in ways that threaten their engagement with school and their academic progress. Adolescents who are highly stressed are particularly prone to both emotional disturbances such as depression and a range of high-risk behaviors such as substance use and abuse. The combination of the initial stress and the mental health or other issue is likely to significantly impair a student’s engagement with school and capacity to keep up academically.

At the same time, disadvantaged children and young people often lack adequate care for medical and mental health problems. Adolescents who have two or more diseases, health conditions, or risky behaviors are particularly vulnerable, and their capacity to succeed academically is likely to be significantly compromised. Adolescents who already have other risk factors—those from low-income families; who are in the foster care system or are homeless; whose families have recently immigrated to the United States; who are lesbian, gay, bisexual, or transgender; and those in the juvenile justice system—tend to have more untreated health problems than their peers. These are young people who will also be likely to arrive at school unready to engage and learn, and to respond to academic setbacks or discipline problems by disengaging. Scholars who have focused on health care for adolescents stress that caregivers, teachers, and other adults who may come into contact with a student or a family are unlikely to have a complete picture of the stresses facing the family, even though the issues are likely to be interrelated and difficult to resolve in isolation.

Families and home environment

A closer look at stressors that can exist within the family illustrates their possible impacts on educational outcomes. For example, physical, sexual, and emotional abuse and neglect may have profound outcomes. Abuse may have neurological or medical impacts on children and young people, including brain damage, neurobiological effects, mental retardation, speech defects, handicaps, and other health problems. Such abuse may also have impacts on cognitive development that manifest themselves in lowered IQ, difficulties with attention, learning disorders, poor reading, poor school performance, or dropping out of school. Possible effects on behavior may include aggression, truancy, delinquency, running away, drug use, and crime and violence. Effects on young people’s psychological states or emotions may include anxiety, depression, dysthymia, low self-esteem, poor coping skills, or hostility.

Although this sort of abuse is an extreme stressor, it is distressingly prevalent. Researchers estimate that about one in seven children between the ages of 2 and 17 are victims of maltreatment in a given year, and that children in low-income families are more likely than others to be maltreated. Low-income children are also more likely than their peers to move (change residence and often school or district) frequently. Not all mobility is harmful in itself, apart from the associated factors that sometimes result in frequent moves, such as family disruption, housing stress, and poverty. Certain kinds of school mobility, however, have a negative effect on children’s educational progress. Negative effects are found in children who move the most frequently, with the impact increasing with each move. The impact of frequent moves is greatest on younger children and on children with other risk factors, such as homelessness. There is a significant relationship between mobility and both lower school achievement and dropping out, and student mobility has a negative impact on schools as well.

These connections are complex to isolate and tend to be obscured in national data, but researchers have found a decline in achievement test scores of approximately one-tenth of a standard deviation for each move a child makes. Dropout rates can increase by as much as 30% for students who have made three or more moves. The effect of high rates of mobility on schools is large. Students moving in and out of classrooms disrupts both the flow of instruction and the establishment of learning communities, and the effects on students of attending schools with high rates of student mobility are likely to persist throughout their school careers.

Children in families who have recently immigrated to the United States or whose first language is not English may also enter school in the United States at a disadvantage (although being bilingual is academically advantageous). English learners and many minority children have lower vocabularies than their peers, a factor that is associated with both lower socioeconomic status and lower achievement. Children who are either not fluent in English or whose home environment is not linguistically rich are likely to have academic difficulties until their language gaps are closed. There is ongoing theoretical debate over the best ways to address the needs of these students, but researchers agree that teachers often lack training in basic strategies for supporting language development.

Involvement in crime and the juvenile justice system is another obvious hindrance to academic progress and school completion. Causal links have not been established, but the importance of interactions among various risk factors is evident. For example, deficits in verbal and reading skills are linked to drug use, aggression, and delinquency. Students who fall behind in reading lose confidence in their capacity to succeed in school, and school failure undermines their engagement. Researchers have shown that young people who experience the familiar sources of disadvantage (prenatal exposure to drugs, low birth weight, and birth trauma; poor language development; abusive or neglectful parenting; and disorganized family and community environments) are those who are more likely to engage in delinquent or criminal behavior.

Delinquency is associated with poor school performance, truancy, and leaving school at a young age. Moreover, research suggests that some educational practices, such as grade retention, tracking, suspension, and expulsion, may intensify students’ disengagement from school. Such practices, which disproportionately affect minority groups, interfere with students’ attachment to school and with learning and unintentionally reinforce negative stereotypes. They can have long-lasting negative effects on academic achievement. One reason researchers in this area point to practices that isolate young people with disciplinary problems is that when those students are together they tend to reinforce one another’s antisocial impulses.

Researchers of delinquency note that the risk factors that make delinquent behaviors more likely tend to be most prevalent in particular neighborhoods, although research on the effects of differences in neighborhoods and their interactions with individual and family conditions is not yet settled. At the same time, students who live in relatively unstable communities in which weapons are readily available and crime is prevalent are more likely to associate with peers who engage in delinquent behavior and to do so themselves.

Program responses

Other countries have focused on reducing the relationship between achievement and family background and have supported disadvantaged students in achieving to high academic standards. A recent report from the Organisation for Economic Cooperation and Development has shown that the highest-performing countries are those that invest in children’s early development and sustain the supports through secondary school. These supports include policies focused on public education as well as other areas, such as housing and welfare. Attention to the links among parents, communities, and schools encourages parents’ involvement in their children’s education, which in turn can make a difference in a range of outcomes.

In the United States, many sorts of programs have shown promise in supporting young people and families who are disadvantaged. These include parenting and home visiting programs designed to improve parenting skills and developmental outcomes for infants and very young children; comprehensive early education programs; interventions for disrupted families; and school-based programs focused on specific goals.

Out–of-home child care and other supports for families can provide opportunities for learning, stable relationships, and other resources that can be critical to the development of children who are at risk. Early Head Start, Head Start, and other high-quality preschool programs are among the interventions that have received a good deal of attention, most recently in President Obama’s 2013 State of the Union address. High-quality early childhood care and education programs that are successful with disadvantaged children share characteristics such as low child-to adult ratios, small group sizes, and a child-centered approach. Programs such as Head Start that focus on school readiness for low-income children provide services that address families’ health, nutrition, and social needs, as well as children’s cognitive and educational needs. They are designed to be responsive to families’ cultural and linguistic backgrounds and may include full-day services and home visits, as well as support and care from health care professionals and others.

Rigorous studies have documented benefits from such programs, including positive effects on cognitive and language skills; achievement test scores; high-school graduation rates; behavior problems, delinquency, and crime; and employment, earnings, and welfare dependency. Reductions in costs for public education, social services, the justice system, and health care have also been found. By one calculation, every dollar invested in high-quality early care and education yields $13.00 in savings to taxpayers. These effects decline over time, in the absence of other supports to sustain benefits, but are strong and remain discernable. For example, high-quality programs may close between 30 and 70% of achievement gaps, depending on how long children receive the services.

Head Start and some other programs attempt to address the multiple needs that struggling families face, but as children age out of the program, their families still experience challenges that may undermine their capacity to meet academic challenges. Many programs are available for older children. Some school-based programs address problem behaviors, such as aggression and substance abuse, and others provide support and training for parents designed to foster connections between school goals and disciplinary and other practices in the home. Many communities have explored possibilities for providing school-based health programs, particularly for adolescents.

Growing recognition of the importance of health and emotional well-being to young people’s academic potential has led many communities to focus on wraparound programs designed to address a range of needs. Wraparound programs have developed in an ad hoc way and are difficult to define. Such programs might encompass parenting, wellness, and other supports, which may be as diverse as dental and mental health care, literacy classes and job training for parents, and extracurricular after school programs for students. These programs take many forms. One example is in California, where in many counties, children and families in need of support (identified through agencies concerned with children and families, mental and health care, probation, and the like) are given a plan of care and support that is modified as their needs change. The key to this program is coordination among county agencies and a commitment to continue the care as long as it is needed.

Wraparound programs are diverse, and research into their outcomes has varied, but summative analysis of the literature has shown the potential for a range of positive impacts on young people and their families for many approaches. Many individual programs, such as the Harlem Children’s Zone, have tracked their own outcomes and show promising results.

Many programs to support families are school-based. The concept of community schools, which, as defined by the U.S. Department of Education provide integrated comprehensive academic, social, and health services to students, their families, and members of the community, has grown in popularity. Community schools are designed to coordinate, for example, engaging instruction, extracurricular learning opportunities, health and social support, and guidance along pathways to college and a career. Advocates for community schools report promising results in improving academic achievement, reducing dropout rates, reducing disciplinary problems, and increasing parents’ involvement in their children’s education.

One approach to integrating services available to families is in North Carolina, where the Departments of Health and Human Services and Department of Public Instruction have collaborated to develop a program of Child and Family Support Teams. In this program, currently serving 21 out of 115 districts, a plan is developed for each student to coordinate supports from family, friends, and neighbors; as well as mental health and medical professionals, juvenile court counselors, social workers, or others who may be needed. Other states and communities (such as Hawaii, Louisiana, Kentucky, and Iowa) have explored ways to support students and families, but the research literature on such programs is limited.

Intersection points

Some of the reports reviewed for this article were early flags of issues that have merited further work, whereas the more recent ones provide syntheses of the most current thinking in various fields. As a group, they demonstrate the magnitude of the resources that are already devoted to understanding and supporting children and adolescents and their families. This brief overview of a wide range of topics demonstrates the complexity of the disadvantages facing many children and young people in the United States. According to the National Center for Children in Poverty, 44% of U.S. children live in low-income families (below $44,700 for a family of four), whereas 22% (approximately 16 million children) live in families whose incomes are below the federal poverty level ($23,021 for a family of four).

Each of the risks discussed here is associated with low income, although they are not confined to young people whose family incomes are below these levels, and the negative effects of risk factors appear to be intensified when more than one is present. What else can be inferred from a review of this work?

• The effects of disadvantage are usually cumulative, so academic difficulties are likely to be symptoms of problems of long standing for students. It seems likely that negative influences from outside of school harm the educational experience, and that negative experiences in school may in turn exacerbate out-of-school problems. Researchers describe dropping out of school as a gradual process, rather than a sudden decision. Understanding of the risks to development in the first years of life provides a good reason to believe that in many cases, the difficulties that ultimately result in a student dropping out could begin that early.

• The interactions among developmental factors, environmental factors, and education practices and experiences may exacerbate harm or mitigate it. That is, there are many possible ways that a child may develop the resilience to thrive despite significant disadvantage, and although early development is critical, intervention that comes later can still be beneficial.

• Integration and coordination of the services available to children and families appear to be essential. A single student’s needs might span the full range of responsibilities that jurisdictions undertake—from health and mental care to the juvenile justice system. Only if these entities are in communication about individual students and their families can they be sure their programs are not working at cross purposes and that a critical aspect of the problem is not being overlooked.

This last point may be the most important. Experiencing multiple risk factors over extended times poses the highest risk to well-being and student achievement. The groups of children and adolescents who are in those circumstances have the greatest need for support that is coordinated across sectors. Any given adult—teacher, public health nurse, or probation officer—might be the one who interacts with a child at a critical point and has the opportunity to identify the need for intervention. It is vital that that individual have the knowledge and resources to quickly direct the student or family to the supports they need, even if the needs fall outside of his or her expertise. A community’s investment in the range of services it provides with the aim of helping young people enter adulthood healthy, educated, and ready to flourish, will be maximized if those services can be coordinated.

Many of the individual reports note the importance of establishing and improving links and communication among the entities that work with young people and families, but references to schools are relatively few. These reports nevertheless throw into high relief the reality that schools are unlikely to narrow the achievement gap on their own. The reality that the greatest achievement gaps are between students with different family income levels has been persuasively documented. These programs and reports demonstrate the potential value in focusing more closely on the specific ways that disadvantages impede the academic progress of low-income students.

Making Fuel Cells Work

For more than five decades, fuel cells have been heralded for their potential as a cost-efficient, environmentally friendly means to convert readily available chemical energy into electric energy. So far this potential has not been realized. In the United States, the foremost reason is that the federal government and the fuel cell industry have pursued a host of research programs that never systematically addressed underlying causes of fuel cell challenges, including reliability, longevity, and cost. Also in this process, the research community, which should have been a key player in this pursuit, has largely been on the sidelines. As a result, the fuel cell industry remains saddled with products that cannot compete with conventional sources of electric power. To set things right, the federal government should implement a national project to support basic research activity dedicated to obtaining detailed knowledge of fuel cell electrochemical operations aimed at solving fuel cell challenges and ultimately attaining commercially viable fuel cell products—at the earliest opportunity.

Fuel cells are singularly remarkable in their potential for efficiently converting the energy locked up in chemical bonds to electrical energy. This efficiency is achieved because fuel cells convert the chemical energy contained in a fuel into electrical energy in a single step, extracting more useful energy from the same amount of fuel than any other known device. An internal combustion engine, for example, requires several steps: converting chemical energy contained in gasoline to thermal energy via combustion, using the thermal energy to expand gases within the engine’s cylinders, and then converting the high-pressure gas to mechanical energy via the pistons and drive train. Because they rely on so many energy conversion steps, internal combustion engines are inherently inefficient and lose much energy to incomplete combustion and exhaust heat. As a result, most internal combustion engines deliver an average efficiency of less than 20%.

Fuel cells can deliver at least twice that efficiency. Moreover, if a fuel cell’s exhaust heat is exploited, total efficiency can be even higher. An internal combustion engine is most efficient under full operational loads (that is why car mileage is lower in city driving) and only at certain minimum engine sizes. A fuel cell, on the other hand, delivers high efficiency even while operating under partial design loads, and its high efficiency is scalable from very small units to very large units.

Fuel cells most commonly use hydrogen to generate electricity, but can use various other fuels, including gasoline, natural gas, ethanol, and methanol. Because of their efficiency, fuel cells using carbon-based fuels create less carbon dioxide and no other emissions, such as nitrogen oxide, that would harm the earth’s atmosphere. Hydrogen presents even a better future. Fuel cells using hydrogen generate no emissions (other than water) and so would have the least environmental impact. Moreover, hydrogen is inexhaustible on earth and can be produced from water through renewable forms of energy— including sun, wind, water, geothermal energy, and biomass—that are readily available in every country. Given their high efficiency, fuels cells, no matter the type, would also slow the depletion of the earth’s fossil fuel resources.

In the long run, fuel cell technology can cause a profound change in quality of life on a global scale. The technology has the potential to give individual citizens the ability to generate clean electricity at home instead of buying it from commercial power plants. Such plants commonly burn coal (and thus emit considerable amounts of carbon and other pollutants), and the electricity they produce is often delivered over great distances by an electrical grid that has high transmission losses and is susceptible to frequent—sometimes catastrophic—failures. Individuals also could drive cars that operate on hydrogen instead of gasoline, thereby dramatically reducing their carbon footprint. Fuel cell technology can spur the development of new classes of energy demand and supply systems, transportation systems, and industrial and manufacturing systems around the globe, while generating trillions of dollars in new revenue. In addition, oil importing nations can cut down on the use of oil, thereby gaining greater independence from oil-producing nations, enhancing their energy security, economic security, environmental security, and the national security.

When fossil fuel is depleted or becomes very costly to obtain, the world will be left with only three energy-conversion options: nuclear fission, nuclear fusion, and renewable energy. The first two options face serious obstacles. Nuclear fission uses uranium, which like oil is a natural resource with high access cost. It also generates long-lived, hazardous radioactive wastes. Nuclear fusion will not be ready for another 40 years. Fusion-powered electricity generation was initially believed to be as achievable as fission power. However, the extreme requirements for continuous reactions and plasma containment have led to projections of feasibility being extended out by several decades. In 2010, more than 60 years after the first attempts, commercial power production is still believed to be unlikely before 2050.

Renewable energy, at least in today’s most common forms, also faces obstacles. Common renewable energy technologies depend on sunshine and wind or biofuels, which are either intermittent, unreliable, or environmentally unfriendly. The best option with renewable energy will be, without doubt, the fuel cell.

Effort, but little progress

The federal government and its partners in industry have long recognized the promise of fuel cells and have worked together to develop fuel cell technology, infrastructure, and manufacturing capabilities. The government alone has devoted billions of dollars in research and development to fuel cell technology in the past 50 years, and corporations have invested similar amounts.

The Department of Energy (DOE) has been in the lead in promoting fuel cell development over the past two decades, spending $2.4 billion on applications research and product development. Most of the effort has focused on developing four types of fuel cells intended for generating power for electric utilities, and on a fifth type, which has received attention more recently, for generating power for vehicles.

But public and private efforts have not delivered on their promise. They have yielded almost no products that can compete with conventional power sources, at least without subsidies and tax credits. Major impediments to fuel cell commercialization include insufficient longevity, reliability, and, in many cases, unacceptably high cost.

Here is a brief summary of these efforts and the results that were achieved:

Alkaline fuel cells. In the 1960s, the National Aeronautics and Space Administration was searching for a source of long-endurance power on manned space flight missions, and decided that alkaline fuel cells (AFCs) would be promising. Rather than try to develop AFCs on its own, the agency asked the United Technologies Corporation to develop them. The company succeeded, and its AFCs were used for the Apollo missions to the moon, the Skylab space station, the Apollo-Soyuz rendezvous and docking mission, and subsequently the Space Shuttle Program. After achieving success in space missions, the company attempted to adapt its technology to terrestrial applications. But it ran into a major problem: AFC performance is easily degraded by very small amounts of carbon dioxide. AFCs used in space missions could be physically isolated from any carbon dioxide generated by the astronauts, and so this was not an issue for these applications. But removing all carbon dioxide from ambient air on earth would be extremely expensive, and so AFCs built for terrestrial applications had short lifetimes and consequent high cost. After considerable effort to solve the problem, the company eventually abandoned the idea. Today, no U.S. companies are engaged in R&D on alkaline fuel cells.

The project should ensure that basic research always be given the highest priority and that it be kept unencumbered from applied research and product development.

Phosphoric acid fuel cells. DOE’s Office of Fossil Energy began funding development projects for phosphoric acid fuel cells (PAFCs) in the mid-1970s, in collaboration with several industry-funded groups, including the Gas Research Institute, the Electric Power Research Institute, and an electric utility consortium. The United Technologies Corporation was the primary contractor. The government and industry undertook this joint effort out of concerns about the cost of controlling emissions from coal-based plants and the rise in cost, complexity, and procedural delays for nuclear installations. After a series of efforts to scale up its PAFC systems to reduce cost, United Technologies began, with DOE’s blessing, to promote its technology for commercial sale. But the high price tag discouraged U.S. electric utilities from making any purchases of the company’s initial 11 megawatt power plant. A typical PAFC power plant was priced at $4,360 per kilowatt hour (kW), and installation costs added an additional $2,000/kW. This was more than six times the nominal $1,000/kW cost for conventional power plants.

Despite the lack of interest from utility companies, the company moved ahead with commercialization of a 200 kW system in the early 1990s. Just as DOE’s support for R&D ended, the federal government began to assist United Technologies by offering a series of subsidy programs that promoted commercialization of the company’s PAFC power plants. These programs remained in place for two decades, but they did not succeed in establishing a going market for the technology. In December 2012, the company sold its power unit (UTC Power) to CleanEdge Power, a small fuel cell company. This effectively ended some 50 years support by United Technologies to promote fuel cell technology—the last major U.S. corporation in the fuel cell industry.

By the time DOE terminated its support, it had spent $300 million on various PAFC power plant development activities. Although the government’s funding assisted the private sector to scale up PAFC technology and demonstrate pilot power plants in the field, the efforts did not solve fuel cell price-performance problems and failed to achieve a level competitive field with other sources of electric power.

Molten carbonate fuel cells. DOE’s support, through its Office of Fossil Energy, for molten carbonate fuel cells (MCFCs) yielded equally depressing results. Funding for initial technology development began in 1975 and continued for 30 years, ultimately reaching $481 million by 2005.

With federal funding, companies were able to scale up their systems and conduct field demonstrations of pilot power plants. But MCFC power plants remained extremely expensive and were sold only with major subsidies and other incentives. Today, only one U.S. manufacturer—Fuel Cell Energy—offers power plants of this type, but the company’s position in the business is increasingly tenuous. Its records indicate that 94% of sales in the United States were concentrated in Connecticut and California, states where subsidies were the most generous. Few other states showed interest. Also, the company depends increasingly on POSCO Energy of South Korea. Sales to POSCO increased from 6.4% of total sales in 2009 to 44% in 2011. Moreover, Fuel Cell Energy has licensed POSCO Energy to use its technology to build power plants in South Korea, eroding the company’s prospects as a manufacturer of MCFCs.

Solid oxide fuel cells. In 1977, DOE’s Office of Fuel Energy signed a five-year, $202 million agreement with Westinghouse— later Siemens Westinghouse Power Corporation— to improve its unique tubular design for solid oxide fuel cells (SOFCs). The design involved forming small hollow tubes with cell components layered on the inside wall of the tubes. The technology proved exceptionally difficult to master, however, and by 2003 lifetimes still were too short and costs were too high to be commercially viable. In 2010, the parent company, Siemens of Germany, put the SOFC business up for sale, after investing more than $1.5 billion and running the company for 10 years.

The project should be backed at the highest level of priority associated with national energy security policy.

Having watched this effort fail, DOE launched a 10-year R&D program in 2000 that focused on new planar designs, where components are assembled in flat stacks one on top of the other as well as more innovative tubular designs for SOFCs. Called the Solid State Energy Conversion Alliance (SECA) program, it was intended to develop an SOFC system that would cost as little as $400/kW by 2010, which was a remarkable goal, considering that the Siemens-Westinghouse designs cost in the range of $17,000 to $26,000/kW.

A critical component of DOE’s plan was to undertake basic research—a step that represented a departure from its previous support for fuel cells. The SECA program instituted a core technology team that would conduct basic research to support the industry teams that would design and fabricate the fuel cells and systems. The core technology team was made up of research entities, including universities, national laboratories, and small businesses, and was to engage in precompetitive research that addressed issues common to all industry teams. The goal was to identify and isolate the underlying causes of fuel cell performance problems in order to drive down costs.

Unfortunately, the SECA program suffered from constant tinkering with its goals. Over a period of five years, SECA was transformed into a program to develop a class of SOFC power plants that would be fueled by coal gas and would produce electricity in the megawatt range; then it was incorporated into a FutureGen initiative to build a fossilfuel power plant that would produce electricity and hydrogen gas without harmful emissions; and transformed yet again into a FutureGen initiative that would use so-called Oxy-Fuel technology to capture carbon dioxide emissions. By the end of the decade, SECA’s fundamental cost reduction goals, which were critical to successful commercialization of SOFCs, had been entirely lost.

One of the more disheartening consequences of these changes was that SECA’s core technology research program lost focus. Although the program received more than $467 million between 2000 and 2010, relatively little went to its basic R&D mission. Consequently, the core technology team did not examine the fundamental principles and mechanisms regarding the physics of fuel cell electrochemistry as it had intended. Rather, it gave priority to “symptomatic relief ” of fuel cell performance problems. Projects became so small in scale—many were under $1 million—that they could not advance fuel cell technology and enable a paradigm shift in cost and longevity. In the end, with lack of funding and commitment by DOE, the SECA program effectively ended in 2012.

Today, the SOFC industry in the United States lacks focus and a clear path for growth. There are about a half dozen SOFC developers in the country; none has publicly disclosed its profitability. Overall, the history of SOFCs is a familiar one: in the absence of a SECA-like project that would systematically support basic research, it is very unlikely that the SOFC industry will achieve breakthroughs in power densities and efficiencies that would yield competitive cost-performance.

Proton exchange membrane fuel cells. Throughout the 1990s, DOE, through its Office of Energy Efficiency and Renewable Energy, supported a small amount of R&D on proton exchange membrane fuel cells (PEMFCs). That changed in 2002 when the Secretary of Energy announced implementation of the FreedomCAR Initiative to develop fuel cell vehicles that would run on hydrogen fuel. In the following year, President George Bush, in his State of the Union Address, announced the Hydrogen Fuel Initiative, which aimed at developing technologies and infrastructure to produce, store, and distribute hydrogen for fuel cell vehicles. The two initiatives were combined into the Freedom-Car and Fuel Partnership.

The partnership ran into implementation problems almost from the start. It inherited the existing organization, the Partnership for a New Generation of Vehicles, that had been formed in 1993 by the federal government and a group called USCAR, an association of Big Three U.S. automakers. This arrangement clearly did not serve the purpose of advancing fuel cell technology R&D. The automakers had commitments to conventional automotive technologies, which discouraged them from focusing on long-term, high-risk fuel cell R&D, particularly if near-term combustion engine technologies were competing for the same funding. So not surprisingly, fuel cell R&D always received a smaller share of funds than hydrogen or conventional vehicle technologies within the total budget.

Reflecting the disadvantaged position of fuel cells in the FreedomCar and Fuel Partnership, the budget devoted to PEMFCs fluctuated constantly. It increased from about $20 million under the original partnership arrangement to as much as $64 million in 2004, declined to $34 million by 2006, only to increase again and decline again in the next few years. Throughout the decade, these fluctuations likely made long-term equitable planning for fuel cell R&D difficult and might have resulted in multiple, insignificant, unrelated, short-term projects. In contrast, the budget for conventional vehicle technologies always received much more funding each year, with no ups and downs. Total funding for the FreedomCAR and Fuel Partnership amounted to $1.6 billion (2002-2012), which was spread among R&D for fuel cells, hydrogen storage technology, advanced combustion and emission control, vehicle systems analysis, and other initiatives. Of the total, fuel cell R&D received $500 million, about 30%.

Was the fuel cell R&D program under the partnership effective? The National Research Council (NRC) reviewed the program and offered mixed findings. In its 2010 report, the review committee spoke well of some of the achievements in the program. It cited an increase in demonstrated lifetimes of fuel cell stacks for on-road vehicles. Indeed, the partnership likely facilitated the entry of some PEMFC products into niche markets, including forklifts and backup power applications. Still, DOE’s 2012 annual progress report stipulated that these products were cost competitive only if subsidies of $3,000/kW were provided and only if they were used for low-power, short duration applications.

The NRC, however, was clearly disappointed that no single fuel cell technology achieved a performance-cost point that would make it competitive with conventional vehicle technologies. In fact, few PEMFCs developed under the program were deployed in the automakers’ own fuel cell vehicles. At the same time, U.S. automakers do not seem to have any near-term plans for new development or commercial launch of fuel cell vehicles.

Why so hard?

What makes fuel cell technology so difficult to master? The answer is simple. Fuel cell operations are complex and the necessary research has not been systematically done to master them. Ironically, this state of affairs is partly due to the fact that it is easy to make a single working fuel cell in the lab, but building fuel cell stacks that generate useful power reliably, efficiently, and cheaply is another matter entirely. History is full of stories in which fuel cell companies developed moderately functional fuel cells, scaled them up for commercial applications (often with government assistance), and rushed the product to market. They subsequently found that their fuel cells did not work well. No matter how many times they encountered problems of cost or efficiency or reliability, however, they always believed things could be fixed with a little more time and a little more money. They were looking for tactical fixes when strategic solutions were needed.

The challenge of making reliable, efficient fuel cells is rooted in the complexities of how they operate, which involves multiple chemical and physical interactions at the atomic level. Perhaps no advanced technology on the market today—including airplanes, computers, or even nuclear reactors—requires the scale, magnitude, and range of scientific, physical, and engineering knowledge that fuel cell technology requires.

Science is making progress, however. Recently, some scientists have successfully applied advanced research tools to understand fuel cell operations at the atomic level and are developing theories to explain details of fuel cell processes. Several examples suggest that this type of work, if integrated into a single collaborative effort, could achieve groundbreaking discoveries that would ultimately lead to viable commercial products:

• Researchers at Tohoku University recently demonstrated how computational chemistry can advance fuel cell development. They linked computationally obtained images with the images obtained from transmission electron microscopy and proved for the first time how computational chemistry can duplicate real-life degradation [of the catalysts? of the electricity-producing materials?] in fuel cells. They have used this discovery to develop a theory of such phenomena, which will help to lay the groundwork to fix the problem.

• Researchers at Yamanashi University and Waseda University, working with others at Shimadzu Corporation, Fuji Electric, and Hitachi, have succeeded in imaging the oxygen distribution within a fuel cell stack for the first time. They used a chemical reagent that absorbs light and emits light of a specific wavelength when oxygen is present and captured an image of oxygen distribution with a charge-coupled device camera. The visualization of the inner working of a fuel cell stack could reveal mechanisms of fuel cell deterioration. They are currently working on the simultaneous visualization of humidity, steam, water, temperature, and carbon dioxide within a fuel cell stack.

• Researchers at Oak Ridge National Laboratory have used a novel microscopy method called electrochemical strain microscopy to successfully examine the dynamics of oxygen reduction/evolution reactions in fuel cell materials, which may reveal ways to redesign or cut the costs of the energy devices. According to the researchers, if they can find a way to understand the operation of the fuel cell on the basic elementary level and determine what will make it work in the most optimum fashion, it would create an entirely new window of opportunity for the development of better materials and devices.

• Researchers at Los Alamos National Laboratory have developed nonprecious-metal catalysts using use carbon, iron, and cobalt to avoid the use of expensive platinum catalysts in hydrogen fuel cells. The team says its next step will be to better understand the mechanism underlying the carbon-iron-cobalt catalyst. Such an understanding could lead to improvements in nonprecious-metal catalysts, further increasing their efficiency and lifespan.

Too valuable to abandon

In considering the disappointing experience with fuel cells, the federal government faces a choice. One option would be to abandon fuel cell development altogether. Another option would be to provide assistance to sustain the temporary survival of the industry. Neither of these options is attractive—or necessary.

Rather, the government can implement what we see as a National Fuel Cell Development Project (NFCDP) that would focus on basic research. Over the past decades, the U.S. federal government, with other governments worldwide, have worked vigorously to help their fuel cell industries build technical capabilities that yielded little economic return. These efforts rested on an assumption that fuel cell technology could be developed into commercially viable products, given enough investment and time to make incremental improvements through applied research and product development. But despite the best of intentions, the fuel cell industry remains in a commercial cul-de-sac and continues to ask for government assistance for survival, including demonstration projects, tax incentives, and subsidies for lowering purchase costs and building a hydrogen infrastructure. Governments are reluctant to stop fuel cell subsidies because they have already invested so much in the industry. And yet, they cannot continue the same policy and subsidy program as before. It is expensive, unproductive, and never ending.

The only viable path ahead is the NFCDP. The federal government should support a basic research strategy and do so as a matter of urgent national priority. Indeed, the NFCDP would focus solely on adding basic research as a key component to the traditional mix of applied research and product development activities that DOE has been implementing for years.

Once fuel cell technology is fully understood and the mechanisms that effect fuel cell performance are known, the fuel cell industry would be able to produce commercially feasible fuel cell prototypes, test them, and acquire solution to every remaining problem within the NFCDP framework. Large-scale simulation, best exemplified by the automotive and aerospace industries, if also implemented, would expedite the commercialization building up from materials and fundamental understanding all the way to manufacturing and distribution of fuel cell products to the marketplace. Thus, the industry could launch commercial products virtually immediately when the project ends, because it already would have a mature technology, well-developed manufacturing capabilities, and marketing infrastructure in place.

The bottom line is that if fuel cell applications are to be successful, basic research must first be successful. Although basic research, applied research, and product development would co-exist in the NFCDP, it should set basic research as its central and primary mission from the start. The project should ensure that basic research always be given the highest priority and that it be kept unencumbered from applied research and product development. Without this protection, basic research will not be able to compete with other research activities; research would likely be directed to abandon its knowledge-seeking mission and shift toward forprofit research that is normally carried out through applied research and product development activities.

At the same time, however, the NFCDP will not be successful unless it links basic scientific research and hypotheses generation to technology development and the empirical data acquired by industry. (This flow is illustrated in Figure 1.) Throughout the life of the project, therefore, the fuel cell industry should have complete access to the NFCDP’s research findings and discoveries and be encouraged to apply these to product development and manufacturing activities. Industry should give the NFCDP continuous feedback on whether fuel cell technology was becoming more durable, efficient, and cost effective. It is essential that industry be fully integrated with the project from start to finish and that companies receive strong incentives to share their information on fuel cell technology and assurance that proprietary data will be protected. The repeated cycle of hypothesis, verification, and data exchange between basic research and applied research and product development will yield improvements in knowledge of fuel cell performance with every step.

FIGURE 1

The Concept of the National Fuel Cell Development Project

What it will take?

The NFCDP’s central and vital mission would necessitate assembling a highly talented cadre of scientists and engineers from all related disciplines in the research, academic, and industry communities. Together they would comprise a critical mass of expertise, knowledge, wisdom, and practical know-how that has never been gathered before for this purpose. The team should be practically and virtually colocated to ensure that team members will be able to integrate their knowledge and collaborate on discoveries in all phases of the project. Most importantly, the team must be led by a strong, can-do manager who has direct access to the nation’s most senior leadership.

The project should be backed at the highest level of priority associated with national energy security policy. Project leaders should be placed directly under the president and be provided with ample resources to explore all possible technology options and momentous freedom. Project leaders also should be granted broad authority to take whatever actions needed, including hiring and firing top leaders and enlisting the nation’s top-quality research cadre.

The NFCDP should have unrestricted access to the national laboratories. They have served as the federal government’s core institutions dedicated to basic R&D and should be deployed as the core of the project’s R&D infrastructure. Importantly, the project must be supported with ample financial resources and extensive research facilities. A research period as long as five years would be desirable to encourage the stability of the research team and continuity of the research environment, while still ensuring that the research team is focused on the tasks at hand. Major technical hurdles can be overcome only through intense, massive investment of resources in a concentrated period.

As a first step toward this goal, a Blue Ribbon committee of top leaders from national research laboratories, industry, and funding agencies should be assembled, under the leadership of the National Academy of Sciences, and charged with preparing a detailed blueprint for implementing the NFCDP.

Among other actions, the committee should identify specific areas of research to be explored. Probable areas of research likely might include the discovery and detailed characterization of fuel cell electrochemical processes and operations; development of a theoretical understanding and empirical validation of underlying causes that drive performance shortfalls, such as cell degradation and insufficient longevity, reliability, and robustness; and exploration of transformational technologies that would enable creation of revolutionary fuel cell types, catalysts, and supporting components using new, less expensive materials. The committee also should identify a range of disciplines that will be needed to support research activities. Research areas likely will include computational chemistry, molecular chemistry, surface chemistry, quantum chemistry, quantum electrodynamics, electromagnetism, mechanical engineering, metallurgical engineering, materials science, mathematics, molecular dynamics, quantum dynamics, thermodynamics, nanotechnology, quantum mechanics, statistical mechanics, atomic and molecular physics, experimental physics, computational physics, nuclear and particle physics, quantum physics, solid-state physics, theoretical physics, and many more.

Of course, there is no guarantee that an effort such as the NFCDP would lead to delivering successful products; it is a necessary, but not sufficient step. Without it, however, there is no reason to believe that fuel cells will make enough progress to become commercially viable. History has proven that to be the case many times over.

Regardless of who would implement the National Fuel Cell Development Project, whether it is the United States or some other country, the world would benefit equally, in terms of a higher standard of living and a cleaner environment. The time to begin is now. Society has in its grasp the scientific and technical know-how to uncover the secrets of the fuel cell—to discover how it works, why it works, and how it can be made to work better.

Shorter Lives, Poorer Health

Americans’ life expectancy and health have improved over the past century but these gains have lagged behind those in other high-income countries. This health disadvantage has occurred even though the United States spends far more per person on health care than any other nation. The United States, which is among the wealthiest nations in the world, is hardly the healthiest. In order to gain a better understanding of this problem, the National Institutes of Health (NIH) asked the National Research Council and the Institute of Medicine to convene a panel of experts to investigate potential reasons for the U.S. health disadvantage and to assess its larger implications. The panel’s findings are detailed in its report, U.S. Health in International Perspective: Shorter Lives, Poorer Health (National Academies Press, 2013). More data, including an interactive chart that makes it possible to see how the United States compares with other countries in the prevalence of specific causes of death such as heart disease, HIV/AIDS, violence, and traffic accidents, can be found at: http://sites.nationalacademies.org/DBASSE/CPOP/US_Health_in_International_ Perspective/index.htm.

Overall life expectancy

The report examines the nature and strength of the evidence on life expectancy and health in the United States, comparing U.S. data with statistics from 16 “peer” countries—other established high-income democracies such as Canada, the United Kingdom, and Japan. For many years Americans have been dying at younger ages than people in almost all other high-income countries, a disadvantage that has been growing for three decades, especially among women.

Set of 17 high income countries ranked by life expectancy at birth, 2007

Lost years of potential life

A measure of years of life lost before age 50 illustrates the amount of premature deaths. The measure shows how many additional years would be lived before age 50 if deaths from a cause were eliminated. Overall, the measure shows that the U.S. loses a larger number of years of life to all disease and injury causes than do any other comparison countries for which comparable data are available. This is true for both males and females younger than age 50.

The causes of extra years of life lost

Deaths from injury—homicide, motor vehicle accidents, and other injuries— contribute more than half of the excess mortality for American males under 50. Noncommunicable diseases are also heavy contributors. For females, injuries also contribute but noncommunicable diseases and perinatal conditions contribute a large part of the excess years of life lost.

Where the United States differs from its peers in cause of death

Closer analysis of the data indicates which causes of death contribute the most to the years of life lost before age 50 in the United States compared to its peer countries. Homicides are particularly prominent for males, whereas noncommunicable diseases and perinatal conditions are more significant for females.

Vested Interests and Geoengineering Research

On March 11, 2011, Japan suffered one of the most devastating earthquakes in its history, followed by a massive tsunami that engulfed reactors at the Fukushima Daiichi nuclear power plant located near the coast. In Japan, the government body that regulates O nuclear power is not highly independent of the utilities it oversees, and regulators had failed to address known safety issues with the reactors. After the crisis, Japan lurched toward a nuclear-free ideology. How can you blame them?

But the catastrophe was not fundamentally caused by a lack of technical information or know-how. Reviewers of the incident found that during the crisis, regulators and company officials made some highly questionable management decisions that were influenced by fears of financial loss and of losing face. Indeed, the catastrophe could have been avoided if good decisions had been made based on available data without the influence of vested interests. As a result, Japan’s energy future has been delimited by human foibles and the resulting breakdown of trust in nuclear energy.

If the tsunami in Japan flooded one coastline and several nuclear reactors, climate change may flood all coastlines and cause worldwide dislocations of people, failures of agriculture, and destruction of industries. The likelihood of these impacts has lent legitimacy to the investigation of intentional climate management, or “geoengineering.” Society may, at some future time, attempt geoengineering in order to stave off the worst, most unbearable effects of climate change. The technical challenge alone is enormous, but Fukushima provides a cautionary tale for managing the endeavor. Is it possible to develop a trustworthy capacity to manage the climate of Earth?

Incentives for manipulation

The potential opportunities, benefits, harms, and risks of geoengineering the climate will almost certainly create incentives to manipulate geoengineering choices, and the stakes will be enormous. Societies globally would be wise to face these potential vested interests as they begin to consider researching geoengineering.

Governments should charter the institutions charged with developing geoengineering research to be rewarded for exposing methods that are bad ideas as well as good.

Vested interests, in this realm, relate to fortune, fear, fame, and fanaticism, and what to do about them. In moderation, seeking fortune or fame, exercising caution, or being guided by philosophy are appropriate and can lead to innovation and good decisions. However, these attributes may become liabilities when nations, institutions, or individuals seek to manipulate the decisionmaking process to make money, enhance stature, save face, or influence decisions based on fanatical ideology. Society can and should expect people to act with honesty and integrity, but should also plan for dealing with vested interests.

Before moving to planning, it is first worthwhile to examine the forces at work.

Fortune. Parties who stand to gain or lose fortunes by promoting or opposing a geoengineering decision have a vested interest in manipulating that decisionmaking processes. Researchers or companies with a financial stake in experiments or possible deployments may seek to push research or deployment in a direction that is ill-advised for society as a whole. Recently, for example, a company desiring to sell carbon credits for sequestering carbon in the ocean conducted a rogue experiment on iron fertilization (the Haida experiment) off the west coast of Canada without obtaining permission or giving due consideration to potential environmental impacts. At this time, there is no legal framework in place to protect society’s interests from a financially motivated company attempting such a geoengineering experiment. In the history of environmental remediation, companies that made money from remediation activities have at times fought changes in regulation that would obviate the need for remediation. For example, California used to require the excavation of soil that had been contaminated by leaking gasoline tanks, until researchers documented that naturally occurring soil bacteria would eventually consume the leaked gasoline, thereby obviating expensive excavation. Companies that stood to make a profit from excavation fought this change in regulation. Similarly, a company with contracts to perform geoengineering would have a vested interest in continued deployment.

Countries that produce fossil fuels and companies comprising the fossil fuel industry may view geoengineering as a way to delay or distract attention from mitigation efforts and thus promote the technology to protect their interests. The chief executive officer of Exxon Corporation, Rex Tillerson, articulated his opinion about climate change, glibly commenting: “. . . we’ll adapt to that. It’s an engineering problem and it has engineering solutions.” The opinion espoused by Tillerson reflects his company’s vested interests. Investigators have documented cases where companies with vested financial interests have bought studies to suppress or manipulate data related to climate change, smoking, and pharmaceuticals in order to obtain favorable opinions, decrease funding, or delay the publication of research. In Merchants of Doubt, Naomi Oreskis, a professor of history and science studies at the University of California, San Diego, described what she saw as the fossil fuel industry’s efforts to manipulate the scientific process and conduct extensive misinformation campaigns related to climate science. These lessons reinforce the idea that the design of a geoengineering enterprise should limit the influence of financial incentives.

Fear. The idea that humans can control the climate is fundamentally hubristic. Individuals involved in geoengineering should be appropriately fearful of this technology and should have great humility and healthy self-doubt that they can control the consequences of intervention.

But there are inappropriate fears that should be avoided. Those involved in geoengineering should not fear losing face when they point out problems or discover negative results. Scientific journals should publish negative results, which for geoengineering are equally as important as positive results. Society surely needs to know if a proposed technology is ineffective or inadvisable.

An institution charged solely with managing geoengineering research would have a vested interest in having geoengineering accepted and deployed, because its continued existence would depend on the approach under consideration being a viable course of action. The institution might be tempted to overstate the benefits of the technologies if it fears losing funding. An institution whose focus is on geoengineering might not want to listen to minority opinions that could slow the momentum of research funding.

As a case in point drawn from recent events, during the economic collapse, there were minority positions within the Bush administration that could have saved the national economy from disaster. For example, the chief of the Commodity Futures Trading Commission, Brooksley Born, repeatedly warned of the dangers of the unregulated derivative market. Her prescient minority voice was suppressed by powerful groupthink within the administration that was vested in economic growth, and she eventually resigned. One cannot help but wonder how many minority voices were suppressed out of fear in light of Born’s experiences. People who are in the minority and sense problems or dangers must not be afraid to speak against the majority or powerful figures who might become invested in the success of geoengineering research.

Fear is a powerful human motivator and often drives institutional culture. It would be a grave mistake to create institutions and power structures in which people are motivated to become overconfident about their ability to control the climate or fear speaking out when they represent minority opinions or are bearers of bad news.

Fame. Perhaps universally, humans have a desire for recognition. Scientists and engineers and other advocates are not immune from wanting to become a Nobel Prize winner, or be called on by the media, or even just have an enviable publication record. The desire for recognition can become a vested interest that leads to a loss of perspective.

Individuals developing geoengineering concepts are likely to know more about the subject than anyone else, and their expertise has tremendous value for society. However, it is always better to have a fresh pair of eyes on a difficult and consequential subject. Society should not depend solely on the developers of technology to assess the effectiveness and advisability of their proposals.

Fanaticism. Unlike climate change, geoengineering is not yet an ideologically polarized partisan issue, but it could become so. Society would clearly benefit from a debate over geoengineering that is grounded in quality information and reasonable dialogue. Fanaticism would polarize and distort the debate and the sound decisionmaking that society requires.

A reasonable ideological position drifts into fanaticism when it hardens into a rigid devotion. Most people have ideological positions on matters of importance, but human philosophies are incomplete and imperfect. For example, in a moment of surprising candor in the aftermath of the 2008 financial crisis, the former chairman of the Federal Reserve Board, Alan Greenspan, famously testified that there was a flaw in his free-market ideology, and that the flaw helped cause the crisis. The tendency to adhere too rigidly to one’s worldview can put one in danger of sliding into fanaticism. Fanatics often use unreasonable and unscrupulous means to promote their causes. To state the tragically obvious: Fanatics can sincerely do much harm.

Many groups and people will oppose the very idea of geoengineering for legitimate philosophical reasons and use honest means to argue against such research. They will raise important issues that need to be debated. However, motivated segments with a vested interest in their ideology or worldview can behave like fanatics, ignoring or misrepresenting factual information and using questionable techniques to create distrust, a situation that could in turn lead to an inability to act strategically in face of climate catastrophes.

On the right side of the political spectrum, for example, an individualistic free-market ideology might lead to fanatical positions that see geoengineering as an alternative to “heavy-handed” government regulations to mitigate greenhouse gases. For example, Larry Bell, an endowed professor at the University of Houston and frequent commenter on energy-related matters, remarked in his latest book, Climate Corruption, that for many on the right, climate change “has little to do with the state of the environment and much to do with shackling capitalism and transforming the American way of life in the interests of global wealth redistribution.” Their vested position, aligned with Exxon’s, could be “we will just engineer our way out of this problem.”

On the left, rigid environmental or antitechnology ideologies might lead some groups to oppose any discussion of geoengineering. Geoengineering is born out of a fundamental concern for global environmental health, but as with the climate problem in general, it has conflicts with an environmental ideology that narrowly focuses on species preservation, regional conservation, and what is called the precautionary principle. (One version of the precautionary principle states: “When an activity raises threats of harm to human health or the environment, precautionary measures should be taken even if some cause and effect relationships are not fully established scientifically.”) Based on an ideology that centers on species preservation, some environmental groups oppose the development of renewable power plants that are designed to help provide energy without emissions but that could cause climate change that could wipe out many species around the globe. In the case of geoengineering, there are environmental groups, such as ETC, that cite the precautionary principle as grounds for banning all geoengineering research, which it sees as a threat to biodiversity. In an ironic twist, rigid antitechnology ideology might become a wedge used by some environmental groups to reject any consideration of geoengineering, even when research is motivated by a desire to preserve biodiversity.

Addressing the four F’s

Moderating the corrupting effects of fortune, fear, fame, and fanaticism should be integral to the development of future geoengineering choices. Society can pay attention to institutional and policy issues that would prevent vested interests from doing harm and provide a counterbalance to human foibles. In this spirit, we offer some guidance for transparency, institutional design, research management, public deliberation, and independent advisory functions. Our suggestions reflect and expand on ideas presented in the Oxford Principles, issued in 2011 by a team of scholars in the United Kingdom as a first effort at producing a code of ethics to guide geoengineering research, and in a report on geoengineering and climate remediation research published in 2011 by the Bipartisan Policy Center, an independent think tank based in Washington, DC. These overlapping strategies each deal with more than one vested interest and could help build genuine trust among scientists, policymakers, and the public.

Transparency. U.S. Supreme Court Justice Louis Brandeis famously said, “Sunlight is the best disinfectant.” When all parties present accurate information clearly and forthrightly, vested interests become less influential. To enable effective public accountability and deliberation, information must be transmitted in a way that is comprehensive, but useful to the lay public. Information users need an accurate understanding of such things as funding priorities, research results, limitations, predictions, plans, and errors. This can be referred to as “functional” transparency, to emphasize that the meaning and significance of the information made transparent should also be transparent, not obfuscated in a blizzard of data.

Functional transparency presents challenges. Scientists use many specialized caveats to express what is known about the climate that are understood by and important for scientists, but can obscure the significance of information and be misleading to nonspecialists. For example, climate scientists have had a difficult time articulating the connections between extreme weather events, such as storms Irene and Sandy, and global climate change. The relationship between weather and climate is complex, and scientists know that such extreme storms have a finite likelihood of occurring with or without climate change. Yet understanding the connections between extreme weather events and climate change is important for effective public deliberations on climate mitigation and adaptation. Scientists’ need for cautious, complex caveats often clouds the issue in public deliberations.

Also, the public is more likely than the scientific community to focus on the context for research. For example, public discourse on geoengineering research, especially outdoor research, is likely to focus on the purpose of the experiment (in fact, some public deliberation never gets beyond this issue), on alternatives to the experiment, on the potential benefits of the methods being researched, and on the potential risks of the experiments being used. This may be especially true whenever there is the possibility that vested interests may be involved, in which case people are wisely concerned about the motives and goals of research. As a case in point drawn from agricultural biotechnology, the debate has largely centered on the motives and goals of research. Proponents of biotechnology often claim that their goals are to address the problems of world hunger and agricultural sustainability. Opponents question these motives, charging that the real goal is a singular focus on increasing the wealth of researchers and their corporate sponsors.

Scientists, however, do not always make the purpose of research completely transparent. For example, some highly legitimate and important climate science research—say, on cloud behavior—simultaneously informs geoengineering concepts. This research spawns the publication of geoengineering analyses, even though geoengineering is not explicitly named as a purpose of the research. Investigators can and do purposely downplay or obfuscate geoengineering as a purpose of research, because this topic is controversial. Just as in the debate about biotechnology, the lack of transparency about the purpose of the research may eventually erode trust and undermine public deliberation about geoengineering. Norms for research transparency should include forthright statements about the purpose of research. There should be clear and understandable assessments of the scope and state of knowledge and expected gains in understanding that could come from research. The transparent release of research information should be designed to inform public deliberation.

As a good example of bridging the divide between scientific discourses and public deliberations, Sweden’s nuclear waste program conducted a study, called a “safety case,” of a proposed repository for nuclear wastes. The study proved a primary tool in developing a public dialogue on the topic, and the process resulted in a publicly approved, licensed facility. The safety case communicated in lay language the technical arguments about why the proposed repository was thought to be safe. It also described the quality of the information used in the argument; that is, how well the factors contributing to safety were understood. The document laid out future plans about what would be done to improve understanding, the expected outcome of these efforts, and how previous efforts to improve understanding performed as expected or not. At a follow-up iteration of the safety case, the results of recent experiments were compared with previously predicted results. Over time, the transparency of this process enabled everyone, including the public, to see that the scientists investigating the future behavior of the proposed repository had an increasingly accurate understanding of its performance.

What lesson does this experience hold for geoengineering? Whereas the goal of this nuclear waste research was to build a successful repository, the goal of geoengineering research is not to successfully deploy geoengineering but rather to provide the best information possible to a decision process about whether to deploy. Nevertheless, the safety case provides a useful example for satisfying the norm of transparency required for effective public deliberations on scientific issues.

There is reason to hope that the propensity for ideological decisionmaking can be limited by transparency in geoengineering research. The experience at Fukushima, however, suggests that the opposite is true: Vested interests can drive nontransparent and poor management decisions that destroy public trust and encourage more extreme, fanatical responses. To engender trust, the people or groups conducting or managing research should explain clearly what they are trying to accomplish, what they know and do not know, and the quality of the information they have. They should reveal intentions, point out vested interests, and admit mistakes, and do all of this in a way that is frank and understandable— all examples of actions that enhance trust. Any subsequent modifications of plans and processes should be transparent and informed by independent assessments of purpose, data, processes, analyses, results, and conclusions.

Institutional design. Institutional design can foster standards of practice and appropriate regulations that will counteract many vested interests. Public funding of research is the first act of research governance, as it implies a public decision to do research in the first place. If research is publicly funded, then democratically elected officials can be held accountable for it. Although public funding would not by itself prevent privately funded research, it would fill a vacuum that private money has so far filled. Furthermore, publicly funded research should not lead to patenting that would produce financial vested interests. Geoengineering should be managed as a public good in the public interest.

Governments should charter the institutions charged with developing geoengineering research to be rewarded for exposing methods that are bad ideas as well as good. One way to obviate institutional vested interest in the success of a method would be to create institutions responsible for a wide spectrum of climate strategies. If an institution investigates an array of alternatives, it would have great freedom to reject inferior choices. In an ideal world, institutions would be created to develop technical strategies for dealing with climate change in general, the defining problem of our time, and these institutions would be given broad purview over mitigation efforts, adaption requirements, and the evaluation of geoengineering.

Just as institutions should not be punished for admitting to failed concepts, individual scientists involved in geoengineering research should not have their careers depend on positive versus negative results. If they discover adverse information, it should be valued appropriately as adding to overall understanding. Organizations that fund research and universities and laboratories that conduct research should publicize and reward research results demonstrating the ineffectiveness, inadvisability, or implausibility of a geoengineering idea. Just as NASA scientists in the early years applauded when a rocket (unmanned, of course) blew up, institutions should reward curiosity and courage in the face of failures.

Research management. Most research in the United States today is “investigator-driven,” in which funding agencies, such as the National Science Foundation, may design a general call for proposals, but the investigators generate the research topics. Funding agencies may convene workshops to explore strategic research needs that subsequently become part of a programmatic call for research proposals. Workshops help to illuminate research that will contribute to an overall goal, but this process does not organize research to achieve a mission per se. There are important previous instances when research with a large-scale public goal was conducted in a collaborative “mission-driven” manner. Now the nation rarely uses this model, and investigator-driven research is the norm.

Geoengineering research (and climate research in general) might benefit from rediscovering, and perhaps reinventing, collaborative mission-driven research modes that focus on a structured investigation of all interconnected parts of the Earth-human-biosphere systems of interest. Interconnections, key failure modes, and critical information needs would be among critical factors to be systematically identified and addressed. The complex, potentially powerful, and intricate problem of intentional management of the climate requires a systems approach. As well, collaborative mission-driven research management would serve to balance the motivation of individuals by rewarding success in meeting the overall goals of research and would compensate for the somewhat random focus of investigator-driven research.

Initial reactions to this suggestion may tend toward the negative, given how mission-driven research was conducted during the Cold War. In the United States and the former Soviet Union, this style of research resulted in massive radioactive releases into the environment, causing extensive contamination of soil, sediments, and surface- and groundwater throughout weapons complexes. Learning from this experience, and using the suggestions offered here, could lead to a reinvention of mission research for geoengineering that would be open, transparent, publically accountable, and environmentally motivated.

There also may be a need to revise the current method, peer review, for assessing the outcomes of research. Peer review will remain necessary—in part, to help balance the exuberance of individual scientists—but by itself will probably be inadequate. Peer review of journal articles would cover geoengineering projects in pieces, without taking into consideration their tight connection to the context and the entirety of the system problem. There is a potentially useful alternative method that was developed for assessing the results of research on nuclear weapons systems that could not be published for security reasons. With severe limitations on access to peer review, laboratories conducting the research pitted two teams against each other to provide checks and scrutiny on research results. In this “red team/blue team” model, one team develops the research, and the other tries to ferret out all the problems. This approach balances a team that might represent institutional and personal vested interests in promoting a technology with a team whose vested interest is in finding out what is wrong with the idea. For evaluating geoengineering research and results, the red team/blue team approach could be considered a more systematic form of peer review.

Public deliberation. Effective public deliberation of the issues, benefits, risks, liabilities, ethics, costs, and other relevant issues will expose and help to neutralize any vested interests that might be in play. Public deliberation can highlight inappropriate profit-making concerns, point out unbalanced scientific positions, call attention to hubris and institutional bias, and counter the influence of partisan positioning on decisionmaking. Public deliberation will be enhanced and facilitated by research that is conducted transparently in trusted institutions that are managed to produce outcomes in the public interest; that is, through all of the suggestions described here. Public deliberation is perhaps one of the few approaches that can help expose ideologies for what they are, whether they come from the political right, which often obfuscates and denies climate science, or from extreme environmentalism, which often uses scare tactics to stop any technological choice anywhere, anytime. No group or individual should get a pass for mendacity in the face of the choices the nation will have to make regarding climate. Public discourse and deliberation will help prevent manipulative dialogue sponsored by ideologues from becoming decisive.

Deliberative dialogue facilitates real-life decisions about setting and prioritizing research goals and selecting the most appropriate means to achieve those goals. For geoengineering, society needs discussions that characterize ethical and social goals; examine competing alternatives; discuss practical obstacles; consider unwanted side effects; assess the technology, including its effectiveness and advisability; and ultimately produce policy recommendations. The deliberative process requires placing scientific research and technological developments in a larger social and ethical context, using this analysis to select intelligent and ethical goals, and identifying appropriate and effective means to achieve those goals.

A recent project in the United Kingdom is a successful example of effective public deliberation on geoengineering. In 2011, a team of researchers planned an experiment to investigate the feasibility of using tethered balloons to release small aerosol particles into the atmosphere that might reflect a few percent of incoming solar radiation and thereby cool things down a bit. The field test would be part of a larger research project, called Stratospheric Particle Injection for Climate Engineering, or SPICE, that involved laboratory and computer analysis of several geoengineering techniques. Although the proposed experiment was nearly riskfree— the plan was to spray a small amount of water into the air—public deliberation about the plans revealed that this work looked like a “dash to deployment” for an immature solar radiation management technology. Research on deployment was not deemed to be necessary or important at this stage. Deliberation also exposed the fact that one of the investigators had intellectual property interests in the balloon technology, and this seemed to violate the principle that geoengineering should be conducted as a public good. Consequently, the investigators themselves stopped the experiment. This honest response helped the investigators accrue credibility and build trust, because their decision responded appropriately to public deliberation.

Independent advisory functions. An independent, broadly based advisory group would facilitate all of these suggested strategies for addressing vested interests. Such a group could help develop standards and norms for transparency, assess and evaluate institutional design, help to develop norms for research management, and lead the way in developing modes and norms of public deliberation. Because the issues raised by geoengineering go far beyond science, an advisory body should also be able to address a variety of broader issues, such as ethics, public deliberation, and international implications. This expanded charge implies that a board’s membership should also go beyond scientific expertise.

Forming independent advisory boards will face a number of barriers. Indeed, the need for an advisory function highlights the inherent controversial nature of geoengineering research, a fact that can make the political choice to start research even more difficult. In the United States, a public board of this type would probably have to meet the standards of the Federal Advisory Committee Act, which is intended to ensure that advisory committees are objective and transparent to the public. Ironically, the act’s requirements effectively inhibit the formation of advisory committees, because they require funding, which is now scarce. There are other potential problems as well. Much current research on geoengineering is very preliminary, and perhaps all of the techniques identified so far will be ineffective, will have unacceptable side effects, or will be impossible to deploy under real-world conditions. Some people in the geoengineering field are concerned that having an advisory board to oversee research might interfere with the research before it has demonstrated that there is anything—positive or negative— that is worthy of oversight. Also, it is not clear what agency or person in the government should form such a board and to whom it should report.

As a practical example of how an advisory body might prove useful, consider again the SPICE balloon/aerosol experiment in the United Kingdom. If there had been an advisory board in place, it might well have recommended that the government cancel the experiment. Such a recommendation would have facilitated government action to stop the experiment, rather than leaving the decision up to the scientists involved, and this step would have given a rather different message to the public about managing controversial research.

The potential value of advisory boards also has been backed up by the research community itself. At a 2011 workshop on geoengineering governance sponsored by the Solar Radiation Management Initiative (an international project supported by the Royal Society in the United Kingdom, the Environmental Defense Fund, and the Third World Academies of Science), participants were asked to consider various forms of organizing geoengineering research. All of them favored a requirement for an independent advisory group, perhaps the only conclusion of this meeting that had unanimous agreement.

In practical terms, consideration of such advisory boards will give rise to many questions about their membership, scope, and authority. To whom should a board report, and how should a national advisory board relate to the international community? How should an advisory board relate to the many governmental and intergovernmental agencies that would almost surely be involved in geoengineering research of one kind or another? Review boards that deal with research involving human subjects cannot actually authorize such research, but they do have the authority to stop research deemed unethical. Should a similar authority be developed for advisory boards on geoengineering research? Answers to these questions should evolve over time, perhaps starting with informal, nonbinding discussions among the various agencies involved.

Just as scientists do not yet know very much about the effectiveness, advisability, and practicality of possible geoengineering technologies, society also does not know very much about how to manage knowledge as it emerges from geoengineering research. If society is to govern this effort without the ill effects of vested interests, it will be necessary to learn how to govern at the same time as researchers are gathering information about the science and engineering of the various concepts. So the early formation of advisory boards or commissions to guide the development of governance is perhaps the first and most important action in countering the potential adverse effects of vested interests and in ensuring that any decisions to pursue or not pursue geoengineering remain legitimate societal choices.

Preparing in advance

Although the future of geoengineering remains uncertain although tantalizing, one thing is clear. It is not too early to begin the conversation about the human weaknesses, vested interests, and frightening possibilities of mismanaging geoengineering. The Fukushima disaster is just one in a long list of reminders of the consequences of not anticipating and moderating the effects of such all too human foibles as fortune, fear, fame, and fanaticism. It is unthinkable that geoengineering should be added to this list of human-caused technological tragedies.

And though much remains to be learned, it is also clear that a number of approaches are already available to moderate the corrupting effects of vested interests: norms for transparency, institutions designed for honest evaluation, management of research in the public interest, public deliberation to expose vested interests and counter fanaticism, and independent advisory boards to highlight and recommend specifics in all of these areas.

The challenge, then, is to get started. Earth is facing ever greater climate threats. Solutions need to be identified and implemented, with all appropriate speed. For many people, geoengineering may offer important help—if the nation, and the world, proceed in a deliberate, thoughtful manner in conducting research and applying the lessons learned.

Buying Whales to Save Them

Since the beginning of the conservation movement in the late 19th century, decisionmakers facing environmental issues have struggled to square the impulse to respect nature’s dignity with more anthropocentric calculations of economic utility. It is a task that has often divided scientists, ethicists, and advocates S who share a regard for biodiversity and ecological integrity, yet differ on how such goals are to be justified and promoted in policy discourse. This debate has evolved over the years as new conservation initiatives and policy proposals have taken center stage, but the core of the dispute remains relatively unchanged: Does viewing species and ecosystems as economic goods preclude seeing them as objects of moral duty? Will the use of economic valuation methods extinguish rather than encourage public support for environmental protection? Can conservation really be expected to succeed by ignoring economic incentives bearing on the protection of wild populations and ecosystems?

A recent proposal to create a “whale conservation market” has highlighted this stubborn ethics/economics divide in a very visible and contentious way. The whale market or “whale shares” idea presents an alternative to the traditional regulatory approach. By calling for the establishment of quotas that could be bought and sold, it allows conservation groups as well as whalers to purchase a fixed number of whale shares, thereby providing a mechanism for whale protection as well as managed harvest. The proposal is for the International Whaling Commission (IWC) to allocate the quotas to member nations on a sustainable-yield basis, which would permit buyers of whale shares to use or sell.

The whale shares idea was first proposed in the January 12, 2012, issue of Nature by one of us (Gerber) and two other researchers: Christopher Costello, the lead author, and Steven Gaines, both at the Bren School of Environmental Science and Management at the University of California, Santa Barbara. The concept was intended to attempt to deal with what many conservationists view as a significant policy failure in international whale management. The IWC, charged with the global conservation and sustainable use of whales, introduced a moratorium on commercial whaling in 1986 as a temporary strategy to conserve depleted whale stocks while a more long-term plan was developed to manage whales. Fueled by interests that challenge the ethics of whaling, however, the ban has not yet been lifted.

Though still in effect, the ban has not been effective. Despite the moratorium, whaling continues at a pace that is widely considered unsustainable. Scientific whaling, which nations (primarily Japan) conduct under the IWC for research purposes, results in the taking of roughly 1,000 whales per year. Subsistence whaling, which the IWC allows for certain aboriginal groups for cultural or nutritional reasons, yields roughly 350 whales per year. Commercial whaling conducted by nations (primarily Norway and Iceland) under objection to the IWC yields roughly 590 whales per year. These harvest totals have been on the rise, as whaling has more than doubled since the early 1990s. The lack of agreement on how to manage whaling despite decades of negotiations between pro- and antiwhaling nations has called into question the future of the IWC as a path to resolution.

Despite the widely acknowledged failure of the IWC moratorium to curtail unsustainable whaling, the whale conservation market idea has proved to be wildly controversial within conservation and antiwhaling circles. Concerns have been raised about how the system would be established (for example, under what guidelines would the original shares be allocated?) and how it would play out over time (for example, would a legal market lead to increased whaling?). Many critics of the idea are also plainly not comfortable with the ethics of putting a price on such iconic species—that is, with using contingent market methods for what they believe should be a categorical ethical obligation to preserve whales.

On the other hand, the negotiation failures surrounding the global management of whales underscore the need for a realistic and pragmatic discussion about available policy alternatives. Indeed, the vulnerable status of many whale populations and the failure of the traditional regulatory response to halt unsustainable harvests call for a more innovative and experimental approach to whale policy, including considering unconventional proposals, such as the whale conservation market. Although it has generated a fair amount of controversy among conservationists, we believe that the whale shares approach does not in fact violate the customary aesthetic, cultural, and scientific regard for whale species; nor does it require the relaxation of the moral commitment to saving species from further decline and slipping into the extinction vortex. But we also believe that conservationists and antiwhaling activists will need to embrace a more experimental policy stance and a less ideological ethical posture if society wishes to make better progress on this intractable international conservation challenge.

Price versus principle in conservation

Presumably, some of the critics of the whale conservation market idea would agree with the great nature advocate John Muir, who warned that “Nothing dollarable is safe, however guarded.” Muir’s skepticism toward economic valuation is fairly typical of the conservation tradition in the United States. The well-known environmental writer Aldo Leopold, for example, wrote in A Sand County Almanac (1949) that society needed to stop viewing the question of good land use as “solely an economic problem” and to do “what is ethically and esthetically right, as well as what is economically expedient.”

More recently, the economics/ethics debate in conservation has centered on the embrace of the ecosystem services framework as an instrument for biodiversity protection. Critics of this move have suggested that the ecosystem services approach fails on moral and practical grounds to ensure the protection of wild species; its supporters have claimed just the opposite. Specifically, to preservationist-minded conservationists, putting a price on nature, or “making conservation pay,” violates the obligation to respect the beauty and moral worth of species and ecosystems. For utilitarian- or economically oriented conservationists, however, it only makes sense to bring nature’s goods and services into the economic sphere, since this is where they can be appropriately valued and traded off against other societal goods. The debate has proved resilient to settlement. Even attempts to broaden the instrumentalist model beyond narrow economic criteria have not mollified critics, who still argue that conservation should be motivated by aesthetic values and a sense of moral duty, rather than by economic considerations.

Although there are important methodological and philosophical issues of environmental valuation at stake here, the debate can often conceal common policy ground among conservationists, clouding the widely held view that the protection of global biodiversity is a primary societal obligation, a duty that exists regardless of whether it is framed anthropocentrically or for the sake of wild species. This contest, especially when it devolves into an ideological struggle between the respective proponents of ethics and economics in conservation decisionmaking, can hinder efforts to experiment with the design of effective institutions to achieve widely supported policy goals. It has a polarizing effect on conservation stakeholders rather than drawing them together in support of new ideas that deserve to be heard and empirically tested.

In the whale conservation market debate, we believe there is an important distinction to be made between, on the one hand, advocating the use of a particular policy instrument (whale shares) to achieve an important conservation objective (maintaining sustainable whale populations) and, on the other hand, arguing that this policy instrument somehow captures the “real” or “full” value of the conservation target. The whale market approach should not be read as presenting an account of the ultimate moral worth of whale populations; it is instead a proposed economic tool that could—we want to emphasize could—help move actors more swiftly toward the policy outcome of effective whale conservation, an outcome that, it is important to note, may be justified by appeal to the shared ethical responsibility to conserve whales and protect global biodiversity.

In his recent book What Money Can’t Buy, Michael Sandel, an influential political philosopher at Harvard University, warns of the “corrosive tendency” of markets in civic life (including in environmental affairs), a malevolence owing to their tendency to “crowd out” nonmarket attitudes and norms, such as the moral and aesthetic regard that society should have for species, by turning such goods into commodities. Sandel lists a variety of examples of this crowding-out process, from commercializing blood donation, to paying children for good grades, to establishing a market to hunt wildlife that simultaneously funnels revenue into species conservation programs. He concludes that by turning nonmarket relationships and goods into market commodities, the character of the good exchanged in the market is transformed as a result of this commoditization process. As a consequence, the conviction to promote the nonmarket good as a moral obligation, such as the commitment to preserve wildlife for its own sake, is eroded.

The argument from moral corruption is indeed an important consideration in these kinds of cases and should be of concern to any conservationist worried about the extinction of moral attitudes promoting a respect for nature. Yet Sandel does not seem to consider a scenario, such as the whale shares idea, whereby a market in wildlife could be designed to also allow conservationists to purchase permits or shares in order to protect (rather than harvest) the wildlife in question. Although one could still argue that a whale conservation market has the potential to crowd out nonmarket values (for example, the love of cetaceans for their own sake), this is by no means an inevitable outcome, especially if concerted efforts are made to promote the moral values of whales as part of the overall management response.

Why whales matter

Discussion of a “shared responsibility” to conserve whales, their “moral value,” however, begs a basic question: Why should society care about the decline of whale populations? For conservationists, the answer may seem self-evident (and perhaps asking it appears unseemly to some). But it is not always obvious how different stakeholders value whales, what ethical responsibilities these values generate, and what tradeoffs they may require in practice. Examining this dimension of the response to the whale market proposal reveals a second, and perhaps more nuanced, division: the distinction between holistic and individualistic conservation ethics.

Focusing on the aggregate value of whale populations, it might be argued that as units of biodiversity they possess intrinsic value, a value derived from their evolutionary history or ecology, or both. This view takes an ecocentric position regarding natural value. It could also be argued on more humanistic grounds that whales (again, considered in the aggregate) should be protected for their unique beauty or cultural resonance, qualities that transcend their utility. Society could value whales for their role in the delivery of ecosystem services, including ecotourism, such as whale watching—a view that recognizes their contributory value for marine primary productivity. Encounters with whales in the wild may also hold transformative value for humans, prompting for some the reevaluation of their material values and preferences. And researchers might make the argument that whales should be saved because of their value for scientific study, including investigations into whale biology, ecology, and behavior; work that could also pay dividends for conservation science and management.

A different range of answers to “why conserve?” would issue from a zoocentric position, which is less concerned with the population- and system-level efforts of most biodiversity conservationists and more focused on the interests or dignity of whales as considered as individuals (that is, as advanced beings that possess aspects of moral personhood, or simply as sentient animals that can suffer). A biocentric view would proceed from similarly individualistic premises, although the moral status of whales is seen to hang less on their advanced mental capacities and more on their status as living beings with biological interests that humans can either subvert or promote. The more vociferous antiwhaling arguments often reflect either a zoocentric or biocentric ethic (or a hybrid of these views) in objecting to what is perceived as the immoral taking of a complex being’s life. The zoocentric/biocentric stance is well illustrated by the recent Rights for Cetaceans initiative spearheaded by the Helsinki Group, a team of scientists and philosophers advocating moral and legal rights for whales and dolphins under the individual “personhood” category.

Although a sustainable use model such as the whale shares approach is not likely to be acceptable under a strict zoocentric or biocentric principle prohibiting harms to individual animals that would occur under any policy of managed use, in principle it could be supported by an ethically diverse array of socially, ecologically, and evolutionarily oriented conservationists who place a premium on the health and long-term viability of biological populations. Yet if the whale shares approach is effective, it would be expected to result in fewer harms to individual whales (by reducing the total take), an outcome that would uphold conservationists’ ethical commitment to sustainable whale populations and zoocentrists’ and biocentrists’ desire to see fewer harms and deaths.

The issue then becomes one of deciding whether the value of ideological purity surrounding the inviolability of animal rights, or of the biological interests of the individual whale, forecloses the pursuit of a policy with the potential to reduce these harms in the aggregate but not end them entirely. By pointing this out, we are not trivializing the difficulty and moral stakes of such choices, but only reinforcing the idea that the international whaling challenge may require conservationists, antiwhaling groups, and animal welfare/rights supporters of a variety of ethical persuasions to come to grips with the tradeoffs and unavoidable political and policy pragmatics of the issue. This means supporting the experimental development of instruments that might be effective in slowing the unsustainable harvest of whale populations, as well as easing the collective suffering of some of the planet’s most remarkable beings.

Sandel argues, and we certainly agree, that market reasoning requires moral reasoning in order to be fair and to keep the former from corrupting the moral norms and attitudes that society attaches to wild species. But his insistence that society will not be able to decide whether to create markets in such goods until it comes to a consensus about the proper way to value them is unrealistic, given that philosophers and others have been arguing over the moral status of nonhuman species for many decades, if not centuries. It is also risky to the extent that it seems to tether any proposal for species conservation employing economic instruments to the precondition of resolving foundational and vexed questions around the moral status of wildlife.

This is not to reject the deeper moral and philosophical project that Sandel describes, but rather to assert that society cannot wait for its resolution before even considering adopting alternative means for protecting species of high conservation value. Although Sandel is right to worry about the extinction of moral norms and attitudes toward species preservation as a result of marketization, he does not worry enough about the extinction of biodiversity and the high stakes of global conservation efforts that may require experimental approaches (including the use of market instruments), approaches joined by the kind of moral analysis and argumentation that he and many environmentalists and antiwhaling advocates champion.

Ethics and experimentalism

A truly principled approach to whale conservation today, and to biodiversity conservation more broadly, requires conservationists to be pragmatic.

The IWC moratorium’s failure to protect some whale populations from decline due to commercial harvest is grounds for a more experimental design for conservation policy. Novel and out-of-the-box solutions, such as the whale conservation market idea, are needed to move society beyond the current policy gridlock. Under this plan, quotas for hunting of whales would be traded in global markets. But again, and unlike most “catch share” programs in fisheries, the whale conservation market would not restrict participation in the market; both pro- and antiwhaling interests could own and trade quotas. The maximum potential harvest for any hunted species in any given year would be established in a conservative manner that ensures sustainability of the marketed species (that is, harvest levels would be established that would not permit taking more individuals than can be replaced) and maintains their functional roles in the ecosystem. The actual harvest, however, would depend on who owns the quotas. Conservation groups, for example, could choose to buy whale shares in order to protect populations that are currently threatened; they could also buy shares to protect populations that are not presently at risk but that conservationists fear might become threatened in the future.

Any future whale conservation market must be securely anchored in a sense of shared responsibility for the conservation of whales and for the protection of globally threatened biodiversity.

Although market-based incentives have helped resolve many environmental challenges, conservation markets still play a relatively minor role in wildlife management. Establishing property rights for environmental goods and allowing trade between resource extractors and resource conservationists may offer a path forward in the global management of whales, one that is superior to the current situation created by the moratorium. Indeed, we and various other observers maintain that such a market could ensure the persistence of imperiled populations while simultaneously improving the welfare of resource harvesters. Although much research is needed before such an approach is implemented, the approach offers a way for both whalers and conservationists to “win” in a workable management regime. Such new ideas are urgently needed, given the failing global moratorium on whale hunting.

At the same time, we believe that it remains imperative to link institutional experimentation with economic policy instruments such as whale shares with the deeper moral commitment to species conservation. That is, any future whale conservation market must be securely anchored in a sense of shared responsibility for the conservation of whales and for the protection of globally threatened biodiversity. Efforts to expand and enhance environmental education and build greater public awareness and appreciation of whales as part of a broader environmental ethic are therefore essential; these efforts will necessarily be part of any principled and sustainable conservation strategy.

Despite the long history of the debate, the use of economic and other policy instruments to achieve conservation goals, and the use of what might be called the “hearts-and-minds” approach to building a sense of moral responsibility for conservation, are not mutually exclusive efforts. We suggest, in fact, that an effective and ethical conservation effort requires that society pursue both types of projects as part of a concurrent, multipronged policy to build institutional capacity and international support for whale conservation. This adoption of an experimental and ethically moored attitude toward whale policy is in step with current strategic and philosophical trends within the wider conservation community to balance moral ideals concerning the protection of global biodiversity with a policy and political realism, as exemplified by the International Union for Conservation of Nature’s Biosphere Ethics Initiative.

There are, however, real challenges facing this task of broadening and informing the conservation ethic. One is educational: The public often does not possess much knowledge about biodiversity loss or the policy agendas and efforts of conservation organizations and governments to respond to these declines. For example, in a study of whale conservation awareness among college students in the United States, a group led by E. C. M. “Chris” Parsons, an internationally recognized expert on cetaceans who currently is based at George Mason University in Virginia, found that the majority of their study sample did not know which whales species were the most threatened, had not heard of the IWC, and did not have a clear grasp of U.S. policy toward commercial and subsistence whaling. These deficits clearly will need to be addressed if the “minds” component of a hearts-andminds campaign is to be successful.

It is also is important to point out that the integrated moral and pragmatic approach to whale conservation described here is squarely in line with one of the most significant traditions in conservation philosophy in the United States. Despite his frustration with economic arguments for conservation, Aldo Leopold, a hero to many biodiversity scientists and advocates, did not wish to purge economic values and tools from conservation policy debates. Rather, he sought to create a more significant space for ethical reflection and argument in environmental decisionmaking and thereby balance utility with beauty and morality in the complex societal calculus of managing and conserving nature.

Leopold implored society to stop thinking about environmental questions as solely an economic concern, not to ignore economic considerations in the formulation of good conservation policy. He challenged society to seek the wise harmonization of economics, ethics, and aesthetics in environmental decisions; a challenge yet to be fully achieved. And Leopold recognized, as today’s society should, that the evaluation of conservation policy proposals requires the experiential testing of ideas rather than a priori judgments driven by dogmatic convictions about the efficacy of certain strategies and policy tools in practice.

Toward principled pragmatism

Whale conservation markets may not achieve Leopold’s ideal integration in conservation policy. But properly developed and implemented, and motivated by a shared commitment to sustaining wild species, they could become important tools for conserving declining whale populations, an activity undertaken for the good of these extraordinary marine species as well as for their contribution to human culture and society. Although uncompromising ethics focused on the interests or rights of whales as individuals have long inspired many antiwhaling activists, the more absolutist versions of these principles can pose a challenge to more pragmatic efforts to experiment with policy instruments that might prove more effective than the IWC moratorium. Regardless, we believe that only experience will be the judge of whether such alternative conservation programs work in practice, and reflective experience will provide the means for improving these policy instruments and strategies over time.

The debate in biodiversity conservation between economics and ethics, or between pragmatism and principle, is in many ways a misguided contest, one that assumes that there exists a deep philosophical division between environmental ethics and societal action. Being pragmatic in whale conservation policy does not mean selling out on conservationist principles.

Rather, we suggest that a truly principled approach to whale conservation today, and to biodiversity conservation more broadly, requires conservationists to be pragmatic. That is, they must be open to policy experimentation where current approaches do not seem to be working, as with international whale conservation, and they must be accommodating of the plurality of interests, values, and methods that increasingly define the wider conservation community. Ethical convictions about society’s duties to sustain whale populations have an important motivational role in any principled and pragmatic conservation policy moving forward.

Needed: Research Guidelines for Solar Radiation Management

Emissions of carbon dioxide (CO2) and other greenhouse gases (GHGs) continue to rise. The effects of climate change are becoming ever more apparent. Yet prospects for reducing global emissions of CO2 by an order of magnitude, as would be needed to reduce threats of climate change, seem more remote than ever.

When emissions of air pollutants, such as sulfur dioxide and oxides of nitrogen, are reduced, improvements occur in a matter of days or weeks, because the gases quickly disappear from the atmosphere. This is not true for GHGs. Once emitted, they remain in the atmosphere for many decades or centuries. As a result, to stabilize atmospheric concentrations, emissions must be dramatically reduced. Further, there is inertia in the earth-ocean system, so the full effects of the emissions that have already occurred have yet to be felt. If the planet is to avoid serious climate change and its largely adverse consequences, global emissions of GHSs will have to fall by 80 to 90% over the next few decades.

Because the world has already lost so much time, and because it does not appear that serious efforts will be made to reduce emissions in the major economies any time soon, interest has been growing in the possibility that warming might be offset by engineering the planet: a concept called geoengineering. The term solar radiation management (SRM) is used to refer to a number of strategies that might be used to increase the fraction of sunlight reflected back into space by just a couple of percentage points in order to offset the temperature increase caused by rising atmospheric concentrations of CO2 and other GHGs. Of these strategies, the one that appears to be most affordable and most capable of being quickly implemented involves injecting small reflective particles into the stratosphere.

There is nothing theoretical about whether SRM could cool the planet. Every time a large volcano explodes and injects tons of material into the stratosphere, Earth’s average temperature drops. When Mount Pinatubo exploded in 1991, the result was a global-scale cooling that averaged about half a degree centigrade for more than a year.

So SRM could work. As undesirable impacts from climate changes mount up, the temptation to engage in SRM will grow. But what if someone tries to do it before we knew if it will work, or what dangers might come with it? The time has come for serious research that can get the world answers before it is too late. To that end, we offer a plan.

Variable effects—and benefits

SRM could be designed to bring average temperatures around the world back to something close to their present levels. But because particles injected into the stratosphere distribute themselves around the planet, it is doubtful whether strategies can be found to cool just some vulnerable region, such as the Arctic. Even with a uniform distribution of particles, the spatial distribution of the temperature reductions will not be uniform. For example, work by Katharine Ricke, then at Carnegie Mellon University and now at the Carnegie Institution for Science, has shown that over many decades the level of SRM that might be optimal for China will move further away from the level that might be optimal for India, although in both cases the regional climates would be closer to today’s climate than they would have been without SRM.

Change in precipitation patterns induced by climate change might present a particularly strong inducement to undertake SRM. But here again, there are some variables and some unknowns. Although the best current estimates suggest that SRM, on average, could probably restore precipitation patterns to approximately those of today, the ability of climate models to predict the details of precipitation is still not very good. Also, some parts of the world are likely to find at least a little bit of warming or other climate change to be beneficial, and so later in this century countries in those regions might not want to return to the climate of the past few centuries, even if they could. In the short term, modest warming and elevated CO2 will probably enhance some agricultural production, although with further warming most agriculture will suffer.

Although SRM could offset future warming, it does nothing to slow the steadily rising atmospheric concentrations of CO2. The higher concentration of CO2 in the atmosphere is already having notable effects on terrestrial and oceanic ecosystems. Some plant species are able to metabolize CO2 much more efficiently than others, giving them a comparative advantage in a high-CO2 world. This is beginning to disrupt and shift the makeup of terrestrial ecosystems.

Over a third of the CO2 that human activities are adding to the atmosphere is being absorbed by the world’s oceans. Today the oceans are roughly 30% more acidic than they were in preindustrial times. Sarah Cooley and colleagues at the Woods Hole Oceanographic Institution have estimated that by late in this century, there will be a dramatic drop in harvest yields of molluscs, resulting in a serious decline in the protein available to low-income coastal populations. Also, acidification is already affecting the ability of many coral species to make reef structures. Many marine experts believe that if emissions and ocean acidification continue to increase, most coral reefs will be gone by the end of this century. In addition to being aesthetically and economically important, reefs (along with coastal mangroves) provide the breeding grounds for many oceanic species and form the base of many oceanic food chains.

Political landscape

Today in the United States, there are many people who doubt that climate change is occurring, or if it is, that those changes result from human action. Congress is no longer pursuing legislation to mandate reduced emissions of GHGs, and many political leaders have been avoiding the issue.

Federal regulatory actions to advance energy efficiency and reduce emissions from coal-fired power plants are making modest contributions to reducing emissions of GHGs, as is the tightening of the Corporate Average Fuel Economy, or CAFE, standards covering vehicles. Indeed, as Dallas Burtraw and Matthew Woerman of Resources for the Future recently observed, these regulations, together with the dramatic growth in the use of natural gas, have placed the United States on a path to achieve President Obama’s goal for reducing U.S. emissions of GHGs. The goal calls for cutting emissions by 2020 to a level that is 17% below levels emitted in 2005. Of course, a 17% reduction does not come close to the U.S. share of reductions needed to stabilize the climate. A few states, most notably California and some in the northeast, are taking direct steps to reduce emissions. Overall, however, the United States shows no signs of being ready to adopt policies to implement the large economy-wide emission reductions necessary to deal with climate change.

Explicit climate policy has progressed further in Europe, where there is a widely shared understanding of the reality of climate change and the risks that it holds. But even as Europe has taken steps to begin reducing emissions of GHGs, these efforts also remain modest when compared with what will be needed to stabilize climate. In the 27 nations that comprise the European Union, per capita CO2 emissions are roughly half those of the United States. However, Europe’s present economic difficulties, together with Germany’s growing dependence on coal as it moves to abandon nuclear power, have resulted in a rate of emissions reduction that now lags that of the United States.

Across the major developing nations—China, India, and Brazil—the primary focus is, of course, on economic growth. China is actively developing wind and solar power, as well as technologies for carbon capture and sequestration. China is doing this because it faces local and regional air pollution that is prematurely killing millions of people, because the government realizes that the country will need to wean itself from coal, and because the government assumes that sooner or later the rest of the world will get serious about reducing emissions and, when that happens, China wants to be a strong player in the international markets.

A tempting quick fix

Although subtle impacts from climate change have been apparent for decades, it is only recently that changes have become more obvious and widespread. Over the next few decades, such changes will become ever more apparent. Because reducing atmospheric concentrations of GHGs is inherently slow and expensive, as more and more people and nations grow concerned, SRM could become a tempting quick fix.

SRM is a technology that has enormous leverage. Recent analysis by a university/industry team of researchers, led by Justin McClellan of the Aurora Flight Science Corporation, suggests that a small fleet of specially designed aircraft could deliver enough mass to the stratosphere in the form of small reflecting particles to offset all of the warming anticipated by the end of this century for a cost of less than $10 billion per year, or roughly one ten-thousandth of today’s global gross domestic product of $70 trillion (in U.S. dollars). In contrast, estimates by the Intergovernmental Panel on Climate Change in its fourth assessment report suggest that the annual cost of controlling emissions of GHGs to a level sufficient to limit warming will be between half a percent and a few percent of global gross domestic product.

Clearly, given this enormous cost difference, as the impacts of warming and other climate change become more apparent, SRM is going to look increasingly tempting to countries and policymakers who face serious adverse impacts from climate change. Adding to this temptation is the fact that implementing SRM could be done unilaterally by any major nation, which is far from the case with reducing global emissions of GHGs, which would require cooperation among a number of sovereign nations around the world.

Planning a research agenda

Although it is well established scientifically that adding fine particles to the stratosphere would, on average, cool Earth, science cannot be at all sure about what else might happen. For example, science cannot be confident about the fate and transport of particles (or precursor materials) once they are injected. It is unknown whether and how the distribution of particles could best be maintained. The surfaces of some types of particles could provide catalytic reaction sites for ozone depletion, but again details are uncertain. Researchers have documented the transient effects of large volcanic injections, but it is not known whether a planned continuous injection of particles might produce large and unanticipated dynamic effects. In short, if the United States or some other actor were to undertake SRM today, it would be “jumping off a cliff ” without knowing much about where it, and the planet, would land. Humans have a long tradition of overconfidence and hubris in considering such matters. In our view, anyone who undertook SRM based on what is known today would be imposing an unacceptably large risk on the entire planet.

The climate science community has been aware of the possibility of performing SRM for decades. However, most researchers have shied away from working in this area, in part because of a concern that the more that is known, the greater the chance that someone will try to do it. Although such concerns may have been valid in the past, we believe that the world has now passed a tipping point. In our view, the risks today of not knowing whether and how SRM might work are greater than any risks associated with performing such research.

We reach this conclusion for two reasons. First, the chances are growing that some major state might choose to embark on such a program. If science has not studied SRM and its consequences before that happens, the rest of the world will not be in a position to engage in informed discourse, or mount vigorous scientifically informed opposition if the risks are seen as too great. Second, given the slow pace at which efforts to abate global emissions of GHGs have been proceeding, the chances are growing that when the world does finally get serious about abatement, the United States and other nations may in fact need to collectively engage in a bit of SRM, if it can be done safely, in order to limit damages, while simultaneously scrambling to reduce emissions rapidly and perhaps also scrub CO2 from the atmosphere.

There have been several calls for a significantly expanded research program on SRM. For example, the House Science Committee and an analogous committee in the UK’s Parliament have explored the issue. The United Kingdom has also undertaken a modest program of research support. A task force of the Bipartisan Policy Center, an independent think tank based in Washington, DC, recently developed recommendations for a program of research by the U.S. government. However, most of the limited research now under way in the United States is occurring as part of existing programs that focus on climate and atmospheric science more generally.

Because SRM could rapidly modify the climate of the entire planet at a very modest cost, and because it holds the potential to have profound impacts on all living things, we believe that there is an urgent need for research to clarify its potential impacts and consequences and to provide sufficient reliable information to enable the establishment of appropriate regulatory controls. Building on the work of the Bipartisan Policy Center, the scientific community needs to develop a robust SRM research agenda and obtain the public and private funding necessary to carry it out. In parallel, the community needs to develop guidelines that ensure that such research is responsibly carried out. Finally, as we discuss in detail below, SRM research should be conducted in an open and transparent manner by providing public notification of proposed field experiments and providing decisionmakers and the public with full access to the results of the research.

Except for limited U.S. authority under the National Weather Modification Reporting Act to require notification and reporting of “weather modification” activities, neither U.S. nor international law provides readily useable authority to prohibit, regulate, or report on the conduct of SRM research or field experiments. Our recommendation is to develop and implement a voluntary research code before attempting to impose any regulatory mandates with respect to SRM research, for two reasons. First, a voluntary code can address and work out the various definitional and policy questions we discuss below. Second, a clumsy U.S. attempt to require notice and reporting of SRM research may simply delay or drive that research abroad, frustrating the ultimate objective of open access to responsibly conducted SRM research. A voluntary code, in our view, is the most sensible first step. The United States should take the lead by developing and implementing a code of best SRM research practices and a set of rules governing federally funded SRM research. After doing that, it should then undertake formal governmental steps and informal steps through scientific channels to urge other international players to promptly do the same.

A key component of a significant SRM research program is to develop a fully articulated research agenda. This might be done under the auspices of the U.S. National Academies, drawing on researchers from major universities, national laboratories, and federal agencies, with input from the international research community.

Code of best practices

In parallel with, or even before, developing a full research agenda, there is a pressing need to develop what we will call a code of best SRM research practices. This code will need three components. The first would comprise guidelines to provide open access to SRM knowledge by making research results available to decisionmakers and the public. The second would be the delineation of categories of field experiments that are unlikely to have adverse impacts on health, safety, or the environment (that is, experiments conducted within an agreed-upon “allowed zone” of experimental parameters and expected effects on the stratosphere.) The third component would be agreement that any field research to be conducted outside the allowed zone will not be undertaken before a clear national and international governance framework has been developed.

The development of this code will require a convening entity and sufficient resources to support activities. Federal funding through an Executive Branch agency might be secured for such an undertaking. For example, Congress could fund a National Research Council study to develop a set of clear definitions and research norms. Perhaps a faster way to get this done would be to persuade a well-respected private foundation to provide the necessary resources. The National Academies or the National Research Council would be appropriate organizations to convene the effort. Alternatively, the American Geophysical Union might do this as part of its recently expanded set of activities in public policy.

Formulating guidelines for SRM research and a policy to advance open access to SRM research must address a set of key issues of definition and scope:

• First we need to define what counts as SRM. Is the technology to be deployed only for SRM, such as a specific type of specially engineered reflective particle, or does it also include multiuse technology, such as high-altitude aircraft designed to deliver mass to the stratosphere but also capable of performing a variety of other missions that are completely unrelated to SRM? To the extent that SRM overlaps with fields of use that do not raise concerns, non-SRM commercial activity might be affected by efforts to single out SRM activity for special attention. What about research on “incidental” SRM? Such current or proposed research might include, for example, geophysical studies of future volcanic eruptions; studies of the atmospheric effects of “black carbon,” the strongly light-absorbing particulate matter emitted by the incomplete combustion of fossil fuels; and studies of the behavior of sulfur dioxide emitted from stationary industrial sources. Any SRM open-access program must define SRM in order, among other things, to minimize its impact on related but uncontroversial commercial activity.

• Next we need to agree on what constitutes SRM research. Does it include theoretical research, literature searches, term papers, and legal memoranda, or should it be limited to experimental research, and if so, does it extend to laboratory research or should it be limited to only field experiments? If the focus is limited to field experiments, how should (and could) basic studies in atmospheric science be differentiated from studies that are more specifically focused on improved understanding of SRM? Trying to make such a demarcation on the basis of experimenters’ intent strikes us as deeply problematic; objective criteria will be needed.

• Activities that should be subject to a requirement of prior notification of SRM research need to be defined. At what stage of a project (planning, approval, or funding) should public notification occur, and in how much detail? Also, what medium or media (for example, a dedicated public Internet site, a Federal Register notice, or a proposal submitted to a designated governmental entity) should be used?

• Any policy respecting public access to SRM research needs to spell out the type of research it covers. Does it cover only completed peer-reviewed research? What about studies whose results are not published or that are in progress but have not reached the stage of publication? What about industry research and abandoned or unsuccessful projects? How can public access be bounded in such a way as to preserve valid commercial interests while providing the appropriate level of public disclosure?

• The allowed zone stipulated for experiments will have to be defined, based on results of existing scientific knowledge. This will include careful delineation of areas of permissible field studies and of a protocol for determining that a proposed field study lies within the allowed zone.

• Finally, there are a series of important policy questions that must be addressed. Should an open-access policy be a voluntary undertaking by researchers? Should the policy be incorporated into federal grants and contracts? Is it feasible to prescribe regulations that require open access to SRM research? Do any of these policies interfere with academic freedom or intellectual property rights?

Importance of open access

As part of the effort to develop a code of best SRM research practices, the United States should develop strategies that ensure that the knowledge developed through SRM research is available to the general public and to national and international policymakers to support informed policy discourse and decisionmaking. The creditability and usefulness of a research program can best be advanced by providing the public with advance notice of SRM field experiments and public access to research results.

The SRM research code of best practices should include a commitment to make public the existence of all SRM research activities, perhaps through a mechanism as simple as posting to a common Web site. It should include an agreement that results from prescribed types of research will be made public (preferably through publication in refereed journals). It should provide guidance on the types of field studies that can be undertaken without any special oversight or approval. And it should express an understanding with respect to privately held intellectual property, as discussed below.

Because most federally funded research would probably already have been described in publicly assessable proposals, posting announcements with an abstract of plans to conduct specific field studies on a common public Web site is not likely to present a significant problem for most investigators. However, asking investigators to post preliminary findings on such a site could be more problematic. This is because some leading journals adopt a strict interpretation with respect to the definition of “prior publication.” We believe that in the interests of promoting open access to SRM knowledge, an effort should be made to induce several top journals to adopt a more lenient policy in the case of work related to SRM.

Action by the U.S. government would set a powerful precedent by a major player in the world economy and world research community.

In developing a voluntary code for research conduct, comment and advice should be sought from federal agencies, universities and other research institutions, and nongovernmental organizations and companies likely to conduct SRM research. To maximize its acceptance, the code should probably draw a line between research results that are to be publicly disclosed and those that do not need to be publicly disclosed so as to protect the commercial interests of technologies with multiple non-SRM uses. The expectation is that once the code is finalized, its recommendation could be incorporated into approval requirements in government and private nonprofit funding arrangements for SRM research and promoted as a model for industrial researchers and non-U.S. researchers.

U.S. government support

Although there has already been some modest support of SRM research from private sources, if a concerted SRM research program is undertaken in the United States, it most likely will involve funding by the federal government as well as some use of the unique capabilities of federal equipment and laboratories. Federal research activities that meet the definition of SRM research should include provisions requiring that an abstract describing the research to be performed be made publicly available upon execution of the underlying agreement. In the case of research involving field experiments, the National Environmental Policy Act may require an Environmental Impact Assessment, unless the proposed project fits into a category excused from such assessment. If an assessment if required and prepared, the public will have ample notice and opportunity for comment.

Federal research agreements should include provisions requiring delivery to the government of publicly releasable research results, commensurate with the SRM research code of best practice. Federal agencies have experience in negotiating lists in each of their research agreements of specifically identified publicly releasable data that would meet the standards set by the SRM code while at the same time, in appropriate agreements, excluding data whose restriction on public release would not be inconsistent with the SRM research code.

Federal research agreements also typically include a patent rights clause that usually provides that the agreement awardee has the option to elect to retain title to its new inventions made under the agreement. In order to lessen the incentive for private commercial interests to influence the direction of the pursuit of SRM, it would be desirable to restrict the assertion of such private intellectual property rights to technical fields other than SRM. Federal agencies already have statutory authority to take prescribed action to restrict or partially restrict the patent rights of awardees. For example, in order to control commercialization, the Department of Energy has provided for federal government ownership of inventions made by its research contractors in the field of uranium enrichment.

A uniform standard can be applied across transactions involving multiple agencies, through mechanisms such as Federal Acquisition Regulations, Office of Management and Budget circulars, and presidential executive orders. Because the promulgation of government-wide guidance may take some time, individual agencies can act on their own initiative if they feel that their mission justifies such action. Individual agency action may lay the groundwork for broader action across the government. If a lead agency is identified to conduct SRM research, that agency should take such an initiative, in the same way that the National Institutes of Health required that investigators who received its support to conduct analysis of genetic variation in a study population submit descriptive information about their studies to a publicly accessible database. The United States could also use international cooperative research agreements as a means to encourage other countries to follow the code of best SRM research practices.

Action by the U.S. government would set a powerful precedent by a major player in the world economy and world research community, giving the nation better standing to advocate for international action in SRM research. Specific U.S. action, developed with input from stakeholders including public interest groups, would establish a model that ensures appropriate public availability of information without unnecessarily affecting commercial interests.

Understand before regulating

The approach we have advocated would have the United States take the lead in developing a set of norms for good research practice for SRM. We have proposed that once developed, these norms should be adopted by federal research programs and urged upon all privately funded research. Once the norms are developed and implemented, it should be possible to persuade others across the international research community to adopt similar norms. Organizations such as the International Council of Scientific Unions and the national academies of science in various countries are well positioned to promote such adoption.

As we noted above, the U.S. National Weather Modification Reporting Act provides a statutory framework for making an SRM open-access research policy mandatory in the United States, at least insofar as the research entails field experiments that are conducted domestically and are of such a scale that they could actually affect climate or weather. Our recommendation, however, is to develop and implement a voluntary research code before attempting to use this authority to implement federal rules governing SRM research.

There is also the question of whether considerations should attempt to go beyond open-access policies for SRM research (that is, notice and reporting) and impose substantive regulation, such as permit requirements or performance or work practice standards. We believe that it is premature today to embark on the development and implementation of substantive regulatory requirements. But as the prospect of large-scale field studies—or actual implementation— of SRM becomes more real, the need for and pressure to develop such regulation will grow. Because future regulations should be based on solid well-developed science, the creation of a serious program of SRM research, combined with procedures to ensure open access to SRM knowledge, is now urgent.

Education, Equity, and the Big Picture

Education has long been recognized as important to individual well-being and the nation’s economic growth. Yet, despite significant public and private investment, disparities in educational opportunities, behavior, attainment, and achievement exist among different student populations. For example, from early childhood through postsecondary education, English learners, low-income, and some racial and ethnic minority students generally have less access to quality learning opportunities, materials, and teachers than their peers. Students from these disadvantaged groups also fare less well on a variety of outcomes. In addition to well-documented gaps in K-12 achievement and graduation rates, these students are more likely to be absent or truant, have disciplinary problems in school, encounter the justice system, and engage in risky behaviors, and less likely to attend and graduate from college. As demographics continue to shift in the United States, the imperative for the K-12 education system to address these disparities and better prepare students to thrive in adulthood is greater than ever before.

Recent policy developments are renewing attention to this imperative and providing an opportunity to address it. The Common Core state standards in mathematics and English language arts will increase academic expectations for all students. In a similar vein, realizing President Obama’s goal of producing 8 million more college graduates by 2020 will mean that a more diverse range of students needs to leave high school prepared for the future.

The available evidence suggests that meeting these goals will require appreciable changes in the approaches school systems take to educate students from low-income families, nonwhite students, and English learners. Making these changes will not be easy, for two major reasons. First, anyone who has spent appreciable time in schools, particularly low-performing schools or those with high concentrations of students in poverty, can attest to the complexity of the educational enterprise. Second, education policies and policy research often belie those known complexities, typically emphasizing one or two levers that are perceived as being the most influential at any given time. Without a greater willingness to grapple with more aspects of the system simultaneously when crafting policies and policy research, progress in reducing educational disparities will continue to be incremental, at best.

In this article I explore the mismatch between the complexity of reducing educational disparities and improving outcomes for disadvantaged students, and the comparatively narrower focus of policy research on those topics (and, by extension, the vision of the policymakers who compel that research). Because the work of the National Research Council (NRC) happens in response to requests from government agencies and other organizations concerned with education policy, the NRC’s portfolio can be used as a lens for reflecting how policymakers view the problem and possible solutions. To inform this article, I examined NRC consensus studies from 1999 to the present that were explicitly focused on or motivated by some aspect of equity in K-12 education. I also examined two recent influential reports on the topic of educational disparities. The first is Equity and Quality in Education, a comparative report by the Organisation for Economic Cooperation and Development (OECD) on some of the policy levers that can be used to reduce system-wide inequities and help to overcome school failure. The second report, For Each and Every Child, was issued by the U.S. Department of Education’s Equity and Excellence Commission in February 2013. That report addresses the disparities in meaningful educational opportunities that give rise to achievement gaps and makes recommendations about the ways in which federal policies could address such disparities.

By highlighting the divergence between the problem and the typical approaches to identifying solutions, this article underscores the need for policymakers and policy research to adopt a broader conceptualization of the problem to identify a wider range of possible solutions for improving outcomes. This article by no means offers a comprehensive review of the literature; rather, the analysis of a few select reports is intended to serve as a starting point for discussions about future directions for research and policy.

The complexity of the problem

In any education system, the major elements of the system that influence the teaching and learning process and educational outcomes include curriculum; instructional practices; school organization and climate; assessment and accountability; state, district, and school-level policies; and finance. The major actors in the system are teachers and their capacity to teach, students, principals, and families.

Each of these elements and actors is conceptually distinct, and each serves a different role in the education system. However, as with any complex system, each element and each actor is connected to the others in a nonlinear way (much like nodes in a web or network), and they are mutually dependent on each other. In addition, the interactions among different actors and between people and the other elements of the system are constantly changing over time, and they are determined by the individuals and their goals for the interactions. The following excerpt from a 2002 article by Jennifer O’Day on complexity, accountability, and school improvement illustrates this point:

In a school, for example, teachers interact with students, with other teachers, with administrators, with parents, and so forth. In each of these interactions, the individual actor follows his or her own goals and strategies, which differ from those of other actors to varying degrees. One teacher may seek to develop her students’ mathematical reasoning while another is more focused on keeping students safe and off the streets. These teachers’ differing goals may be manifested in somewhat different strategies in the classroom. Similarly, while one student may show up because he wants to hang out with friends and avoid the truancy officer, another has his eyes on an elite university. Again, students’ particular activities will vary according to their goals.

Those interactions also are governed by the constraints that the system imposes, such as the length and structure of the school day, state or district standards, the standardized testing schedule, and the curriculum materials and other instructional tools that are available to teachers.

Layering in the needs of disadvantaged, low-performing students and schools increases the complexity of the education process. Students’ background characteristics and conditions typically determine the schools they attend, and the neighborhood context, in turn, determines the availability of resources and the capacity of the professionals who work at the school. Indeed, English learners and students living in poverty require more intensive resources to support their learning, and many schools in poor neighborhoods lack the human and material resources to “break the nexus between student background characteristics and student achievement,” as described in the NRC report Making Money Matter. Paradoxically, the schools that need the highest-quality teachers the most—schools with high concentrations of low-income or low-achieving students—have the hardest time attracting them.

Some constraints or elements of the K-12 education system disproportionately affect disadvantaged students and thereby complicate the process of improving their educational outcomes. As one example, tracking, or the practice of sorting students into different courses or curricula based on ability, has many deleterious effects. Students in lowertracked courses often are low-income, nonwhite, or English learners. Tracking in the early grades precludes these students from taking more rigorous courses over the course of their academic careers, isolates them from their higher-achieving peers, reinforces low academic expectations, and leads many to believe that they lack academic competence. By secondary school, students in these lower tracks might disengage and eventually drop out of school. Nonwhite students also are disproportionally represented in special education and subject to discipline policies that remove them from the school or classroom, further limiting their learning opportunities.

Students’ background characteristics and conditions also influence their own education in well-documented ways (see the companion article by Alexandra Beatty for a more detailed discussion of the out-of-school factors and needs that affect K-12 behavior, achievement, and attainment). For low-income students, lack of access to quality early learning opportunities can limit the extent to which they enter kindergarten ready to learn, shaping the course of their education for years. Being hungry, chronically ill from an absence of health care, or stressed from exposure to hardship or violence can diminish concentration, engagement, and performance in school. Older students who take jobs to support their families might attend school sporadically, which affects their education and the functioning of their classes. Improving outcomes and reducing disparities for disadvantaged students is even further complicated by the ways in which students’ lives and out-of-school influences affect the functioning of the school as a whole, and these dynamics are less well understood from a research perspective.

Policy levers

Looking across the research related to the topics in the previous section and in Beatty’s article in this issue, what we know about improving educational outcomes for disadvantaged students suggests that the process begins in early childhood, requires attention to psychosocial development as well as academic growth, and mitigates the effects of circumstances and conditions, such as poverty, that hinder students’ learning. Most research that is driven by policy concerns and designed to inform policy, however, does not reflect this broad view, and is therefore limited in its utility to help policymakers consider and address the problem in all its complexity.

A review of NRC consensus reports in the past 15 years that have been driven by some aspect of equity in K-12 education illustrates a striking shift in national policy priorities. In 1999 and 2000, three consensus reports on K-12 education focused explicitly on equity. In 2001, Congress authorized the No Child Left Behind Act (originally the Elementary and Secondary Education Act), which included strict accountability requirements for all students. From 2001 to mid-2012, as No Child Left Behind came to dominate the education landscape, just five NRC consensus reports were driven by the goal of equity in education. Most of the reports fall into two broad topic areas: 1) high-stakes testing and accountability and 2) special populations of students, such as dropouts, minority students, students with disabilities, and English learners, typically as they intersected with testing and accountability.

Reports on each of these topics, however deep and comprehensive, have addressed only a slice of the problem of reducing educational disparities for disadvantaged students. The focus of each report also reflects national educational policy priorities at the time when the report was commissioned. Of course, it is valuable to have a deep understanding of individual elements of the system, but homing in on one aspect in this way does not give policymakers a sense of the limitations of approaching the problem one lever at a time. If it is clear that the problem involves complex interactions among the actors and elements of the system, policymakers must grapple with those complexities. This might explain why many NRC reports caution against the search for a magic bullet and point to the need to attend to other elements of the system, even if they do not address those other elements themselves.

Notably, three older NRC reports—written before No Child Left Behind—adopted more comprehensive approaches to the challenge of promoting greater educational opportunity and improving outcomes for disadvantaged students. The committee that wrote the 1999 report Making Money Matter was charged to answer the question, “How can education finance systems be designed to ensure that all students achieve high levels of learning and that education funds are raised and used in the most efficient and effective manner possible?” In addressing its charge, the committee observed that:

Education finance is only one part of a total system of education. Many of the concerns about the financing of education reflect large issues regarding the overall education system. Hence, proposals for changing the finance system can be presented in at least two ways: (1) as a menu of options for driving the education system in desirable directions or (2) as intertwined components necessary to achieve a given vision of overall education reform.

Thus, Making Money Matter tackled broader issues such as the need for investing in the capacity of the system and professionals in the system to improve achievement, and the need for systematic inquiry into a range of more comprehensive and aggressive reforms in urban schools to improve educational outcomes for disadvantaged students.

Initially driven by concerns that high-school exit examinations could have the unintended effect of increasing dropout rates among students whose rates are already far higher than the average, the NRC report Understanding Dropouts (2001) examined the state of knowledge about influences on dropout behavior and the available data on dropouts and school completion. By definition, that examination probed the interaction of nonschool factors and many different aspects of the education system and their role in dropout.

In a related vein, the 2003 report Engaging Schools also had a relatively broad charge to examine what was known about promoting disadvantaged urban adolescents’ engagement and motivation (and academic achievement) in high school. That committee looked at the intersection of several different aspects of the education system (curriculum, teachers, instructional strategies, and the organization of schools) with broader research on motivation and engagement. The committee also took into account the role of family, peer culture, and community resources. Reflecting this comprehensive view of the problem, the report’s recommendations range from greater use of classroom-based assessment of students’ knowledge and skills to restructuring urban high schools to improving communication, coordination, and trust among the adults in the various settings where adolescents spend their time.

Other approaches

Moving beyond the NRC portfolio, two recent policy reports provide contrasting approaches to their recommendations for reducing disparities. The 2012 OECD report Equity and Quality in Education is a cross-national examination of the approaches governments can take to prevent school failure and reduce dropout. The report notes that

Governments can prevent school failure and reduce dropout using two parallel approaches: eliminating system level practices that hinder equity; and targeting low performing disadvantaged schools. But education policies need to be aligned with other government policies, such as housing or welfare, to ensure student success.

At the system level, the OECD report suggests that the following policies could help to reduce or avoid creating disparities in educational outcomes:

  • Eliminate grade repetition
  • Eliminate early tracking
  • Make funding strategies responsive to schools’ and students’ needs
  • Provide flexible pathways for students to complete secondary education

In terms of supporting low-performing schools and students to improve, OECD’s recommended strategies include:

  • Attracting, supporting, and retaining high-quality teachers
  • Creating a school climate that is conducive to learning
  • Strengthening school leadership
  • Promoting the use of effective classroom learning strategies
  • Prioritizing links between schools and communities

These recommendations are consistent with the major elements of the K-12 education system I described in the previous section and that are commonly identified as the “usual suspects” in education policy reports and conversations. And if implemented in U.S. schools, they probably would improve the process of schooling, and perhaps even educational outcomes, for low-income students, nonwhite students, and English learners. But by how much?

The OECD recommendations address a few levers or (staying with the image of the education system as an interconnected web) nodes of the system at a time and do not fully engage with the interactions among them or with the multifaceted causes of disparities in education. To be fair, the report does acknowledge some of the interrelations among the levers, and in the excerpt above, recognizes that the causes of disparities in educational outcomes extend well beyond the education system. But breaking the recommendations into such discrete pieces could tempt policymakers to treat the recommendations as a disconnected set of menu options and might lead them to expect more progress than is reasonable from selecting one or two items from the menu. Without a richer, more fully developed conceptualization of the problem, clear linkages between the solutions and the nature and magnitude of the problem, and an exploration of the tradeoffs involved in tackling one or two levers at a time, the kind of change that is needed to make appreciable progress in reducing educational disparities will probably remain elusive.

For Each and Every Child, the 2013 report from the U.S. Department of Education’s Equity and Excellence Commission, offers just such a comprehensive conceptualization, together with a set of interconnected solutions that are directly related to that conceptualization. The Equity Commission’s report contains a “five-part framework of tightly interrelated recommendations to guide policymaking.” The parts are as follows:

  • Equitable school finance systems
  • Teachers, principals, and curricula that give students an opportunity to thrive
  • Early childhood education with an academic focus
  • Mitigating poverty’s effects by providing access to early childhood education and a range of support services
  • Reforming accountability and governance systems

This framework (and the report) begins to embrace the complexity of reducing educational disparities by explicitly including early childhood education and acknowledging the effects of poverty on educational behavior, achievement, and attainment. And although finance, accountability, and governance arguably represent a few individual levers or nodes of the system, the report makes a compelling case for the ways in which they affect all educational decisions. Moreover, the tightly interrelated nature of the recommendations— linking, for example, suspensions and expulsions that remove disadvantaged students from the learning process to the lack of funding and disparities in high-quality teachers who can provide supports for them to catch up or stay on track academically—lends credence to the commission’s call for more coherent and coordinated policy action at the federal, state, and district levels. Perhaps the Equity Commission’s report will inspire policymakers to approach this challenge of improving educational outcomes for disadvantaged students in the more comprehensive and integrated way that it deserves.

A downside to the comprehensive nature of For Each and Every Child is that, on the surface, the commission could be accused of playing it safe by being so comprehensive and including such a wide range of solutions. In fact, reforming governance and accountability structures in the U.S. education system or expanding the reach of schools to mitigate the effects of poverty are anything but safe! Still, faced with such an array of potential, and potentially doable, actions, policymakers could benefit from guidance to avoid a linear, lever-by-lever approach or treating the report’s recommendations as a to-do list rather than the long-term interconnected agenda that it is.

A comprehensive view

As I have mentioned, robust bodies of research, including numerous synthesis reports by various NRC committees, exist on most of the elements and actors of the education system as they relate to reducing disparities for disadvantaged students. The commonly identified levers—funding, standards and accountability, teachers and school leaders, instructional practices, school climate, policy—are undoubtedly important to learning in their own right and as interconnected parts of the system. However, as a nation, our approach to moving these levers to improve educational outcomes for disadvantaged students has been ineffective. Moving forward, how can we avoid policy recommendations that read like laundry lists of potentially disconnected strategies that are not commensurate with the complexity of reducing educational disparities and improving outcomes?

Although simultaneously addressing all of the factors that influence educational achievement is undoubtedly difficult, and perhaps even infeasible, focusing on just a few policy levers or on just one youth-serving sector constrains the options available to improve educational outcomes for disadvantaged students. And while the education system cannot be expected to solve broader economic and societal problems that influence educational outcomes, strategic policies might support schools and districts in taking more deliberate steps to identify and mitigate their effects to the extent possible in the course of educating students.

Policymakers need clearer guidance from research to make these decisions, particularly about the limits of one-dimensional approaches that do not address the interconnectedness of the elements and actors in the system. Instead of asking which aspects of the system offer the highest-leverage points for improvement, a more productive line of questioning might be, “How much progress are we likely to make by focusing on just one or two elements of the system? If we did focus only on one lever, how long it would realistically take to realize the desired goals? And at what point would it become necessary to begin grappling with improvements to other elements of the system?” To provide an example, a current policy focus is on evaluating teachers’ effectiveness. The questions above become: How much progress will we make in improving educational outcomes for disadvantaged students by focusing only on teachers’ effectiveness (and, as is often the case, on policies to remove the lowest-performing teachers)? If we did focus only on evaluating teachers’ effectiveness, how long would it take to reduce achievement gaps and improve outcomes for disadvantaged students to the desired levels? And at what point will it become necessary to begin grappling with improvements to other elements of the system that influence teachers’ performance, such as their training and professional development, the organization and climate of their schools, the high turnover among their principals, or the nonacademic needs of the students they teach?

As I have discussed, a few examples of policy research exist that might encourage this kind of thinking. Specifically, in this article I focused on three NRC reports between 1999 and 2003 that undertook relatively comprehensive examinations of reducing disparities and improving educational outcomes for disadvantaged students. In different ways, each of these reports reflected the complexity of that challenge. The far more recent report For Each and Every Child by the U.S. Department of Education’s Equity and Excellence Commission offers an excellent framework for considering multiple aspects of the education system and how they relate to each other. It also addresses the complexity of reducing disparities in educational outcomes by looking beyond the confines of the K-12 education system for influences on achievement, and at reasonable actions for the school system to mitigate those effects. That report also should be commended for calling attention to the long-term nature of implementing such an ambitious agenda.

On the whole, though, policy research contributes to, or at least does not seriously challenge, the one-lever-at-a-time approach. Despite what we know about improving educational outcomes for disadvantaged students, very few individual reports on this topic span the years of early childhood through early adulthood, consider multiple (much less all) elements of the system at once, consider ways to leverage successful interventions from all of the sectors responsible for aspects of young people’s well-being, or involve rethinking the education system in ways that integrate these multiple dimensions to improve achievement and related educational outcomes for disadvantaged students. Yet, given our nation’s dismal track record in educating disadvantaged students, this broader vision of what it means to effectively educate students is exactly what is needed.

To identify directions for future research and policy, I return again to the vision of the education system as a web of interconnected nodes. Although the field has a relatively solid understanding of most of the individual nodes in their own right, research is still needed on certain aspects of individual nodes and on the role of the individual nodes in the larger web, particularly as they relate to reducing educational disparities. Research is also needed on the threads between different nodes, to understand which nodes are more strongly connected to each other, to the teaching and learning process, and to student outcomes. Finally, research is sorely needed on the web as a whole, either through new studies of the education system or by looking across existing bodies of research on different elements of the system. Enhancing the policy understanding in these ways might be a promising start toward the broader vision and ambitious education policies that our most disadvantaged students deserve to help them thrive in the education system and beyond.

Applying the Social and Behavioral Sciences to Policy and Practice

Smart electric meters and the smart grid are innovations in the delivery of electrical services to homes and businesses that can, in principle, permit both consumers and suppliers to save energy, exercise greater control over the uses of energy, and in some cases feed locally generated electricity back into the grid. Europe is adopting this technology, which can play a vital role in achieving environmental goals. In the United States, however, there is resistance prompted by understandable concerns about data privacy and physical security as well as fears about the health effects of wireless telecommunication. Behavioral research has helped overcome this resistance through new methods of data collection and distribution that protect individual privacy while preserving the energy-saving and informational value of smart metering.

Helping people and their technologies work well together is only one of many ways in which behavioral and social sciences contribute to national well-being. Researchers in these sciences have contributed to public policies in areas ranging from defense and national security to health care and education. However, these contributions are not always apparent to the broader science community and the public or in policy settings.

This lack of visibility risks underutilizing the nation’s scientific capabilities as new challenges emerge. For example, as the elderly grow in their proportion of the population — a demographic trajectory we can foresee from surveys designed by social scientists—there will be demand for changes in work arrangements, retirement systems, and health care services, as well as in the design of transportation systems, homes, and neighborhoods. Effective, efficient, and satisfying accommodations to these changes can benefit from the application of the tools of social and behavioral science.

There are also many examples of contributions where the social and behavioral sciences intersect with the physical sciences and engineering. When there is a bridge to be built, we turn to engineers; when there is a vaccine needed, we turn to biochemists. But in determining the relative efficiency of paying for the bridge with a toll or a tax, we turn to economists; and in understanding public acceptance of the vaccine, we turn to social psychologists and public health experts. Society has many challenges that require attention by engineering and the physical and biological sciences, a large number of which also have a human and social dimension.

It is within the reach of political science, decision theory, cognitive psychology, behavioral economics, and systems theory to focus analytically on the “use of science as evidence in public policy.”

In recognition of the role of sciences that investigate these dimensions, the National Research Council has launched a new initiative, Social and Behavioral Sciences in Action (SBSIA). This initiative presents key contributions that the social and behavioral sciences have made to policy and society; it will also highlight emerging areas requiring fresh attention. This article, which draws from presentations at the first SBSIA symposium in September 2012, provides details about past successes and future challenges.

Better health care

One of the most socially beneficial but often unnoticed roles the social and behavioral sciences play is facilitating health care research and practice. For example, demographers and other social scientists were instrumental in developing and introducing modern family planning methods and practices around the globe; for example, by illuminating the key role social networks play in people’s decisions to use contraception. Dartmouth’s Atlas of Health Care has collected, analyzed, and publicized geographic variations in health care expenditures in the United States. Atlas researchers found that the rate of hip replacements was four times higher in some U.S. regions than in others and that the rate of shoulder replacements was 10 times higher in some regions. Such data helped policymakers and health care leaders identify opportunities to reduce health care spending while improving the quality of care.

Cultural as well as economic barriers impede health improvements. Rita Colwell, former director of the U.S. National Science Foundation, led a three-year study to evaluate a new method for combating cholera in Bangladesh. That nation faces cholera epidemics every spring and fall, and the disease spreads when people drink contaminated water. Colwell’s team sought to evaluate whether villagers could lower the incidence of cholera by pouring their drinking and cooking water through old folded sari cloth to filter out contaminants. With the help of statisticians and social scientists, they conducted a study that involved 150,000 individuals in 50 villages, including a comparison group that did not use the sari filters. The researchers found that the group that used the filters decreased their rate of cholera by 50%, and that those who did develop the disease generally got milder cases.

The social sciences’ important role became clear early in the study, when they helped the researchers scale an initial hurdle, said Colwell. When the team first submitted the proposal for the study, one of the reviewers rejected it, saying that Indian men would never drink water that had been filtered through old “unclean” sari cloth. The research team, however, found that in fact the men were already using sari cloth to filter flies out of their beer. “We really needed to understand the cultural practices,” explained Colwell.

Sociologists guided the team’s introduction into local communities, advising them on how to present the study. The sociologists also helped design the questionnaire, ensuring that specific questions were framed in culturally acceptable ways. What would have happened if the research team had not included social scientists? “I wouldn’t have had the entre to the villages on such a grand scale—150,000 individuals in 50 villages,” said Colwell. “It would not have been possible. And it would’ve been tragic, because this is an opportunity to take very advanced technology, science, engineering, and [use] those findings in a very practical way to help people.”

An area of health care where the social and behavioral sciences are badly needed but have not yet been tapped is in reforming the culture of medical practice, according to Lucian Leape of the Harvard School of Public Health. In particular, Leape argued, a dysfunctional medical culture is undermining patient safety; a field that for many in the profession really began with the Institute of Medicine’s (IOM) 1999 report To Err is Human. Before that report, patient safety was seen as an individual performance issue. “If you didn’t perform well, then it was because you were lazy or careless or ignorant, and if you made mistakes we would punish you,” said Leape. “The Institute of Medicine said, it’s time to change the paradigm, to recognize that errors are caused by bad systems, not bad people, and let’s get to work changing the systems.”

Although considerable work has been done since, with progress in fits and starts, it is hard to prove that efforts to reduce errors have had more than marginal impact, said Leape. It’s apparent that changing systems is very difficult, he continued, and that health care is not a learning culture. Medical education emphasizes individual performance rather than teamwork, and medical culture is hierarchical and siloed. “Changing that culture has got to be the ultimate social science challenge,” he said. For example, a study in eight hospitals around the world found that the use of surgical checklists reduced complications by 40% and deaths by 50%. But it has been hard to achieve these good results on a broader scale; in every hospital there is at least one physician who resists such checklists, feeling individually exempt. ”The simple way to say it is that we don’t do teams very well,” said Leape, “and that’s clearly a social science issue.”

Of course, we also know that the origins of many health conditions are social in nature. In explanations of premature death in the United States, observes Steven Schroeder, social, behavioral, and environmental factors far outweigh genetic factors or problems within the health care system. A 2013 report of the National Research Council (NRC) and IOM has shown that the United States lags well behind other developed nations in morbidity and mortality in almost every respect for individuals less than 75 years old and that social, environmental, and behavioral factors are largely responsible for this disparity. Schroeder found that 60% of premature deaths are attributable to social circumstances, environmental exposures, or behavior patterns, whereas 30% are due to genetic defects and only 10% to the health care system. Innovative longitudinal biosocial surveys of health and aging in the United States and other nations promise to add substantially to knowledge about social and behavioral factors in health and longevity.

Enhanced national security and public safety

Another context in which a better understanding of human behavior can save lives is presented in the work of national security psychologist Robert Fein, who has studied the preattack behavior of assassins and school shooters. In the late 1980s, after four cases in which individuals had tried to attack someone under Secret Service protection, the agency’s director asked Fein and a Secret Service agent to conduct a study that would provide relevant evidence to Secret Service agents about would-be attackers. The researchers examined the pre-attack thinking and behavior of 83 people who had tried to attack a public target.

What they found debunked some commonly held myths about would-be assassins, especially the misleading assumption that there is a standard profile of an assassin. In fact, attackers vary by age, personal background, and other characteristics. Another common myth is that mental illness is a direct cause of assassination attempts. In reality it rarely plays a critical role; when agents try to assess a threat, it may be far more important to ask how organized a suspect is than to ask whether he or she is mentally ill, the researchers concluded. They also found that those who pose threats may not make threats. “No presidential attacker has ever communicated a direct threat to the president, to law enforcement, or to the Secret Service,” said Fein. His findings were disseminated to Secret Service agents and law enforcement agencies around the country and led to changes in investigative activities designed to prevent attacks.

After several high-profile school shootings rattled the nation in the 1990s, the Secretary of Education asked the Secret Service to conduct a similar study of the shooters’ pre-attack behavior, and the agency once again turned to Fein and his colleagues. Using an approach similar to their study on assassins, they examined the thinking and behavior of 41 young people involved in attacks at school. As with the assassins, they found that there is no accurate, useful profile that can predict school violence. However, they did find that most attackers had come to the attention of adults before the attack for problems they were having;plus, many attackers felt bullied, persecuted, or injured by others. In addition, most attackers let others—usually other kids—know that they were planning attacks, but those who knew rarely told adults. These and other findings had a broad impact when Fein and his colleagues developed a series of video teaching tapes and presented them at meetings of school and law enforcement professionals across the country.

Many other areas of the justice system and national security policy have benefited from greater application of behavioral and social science research. For example, two 2011 NRC reports showed that the adoption of behavioral and social science findings and methods could improve the quality and effective use of national intelligence analysis. Another NRC report demonstrates that widely used industrial methods of quality control could speed development, reduce costs, and increase the effectiveness of new defense systems. Experimental and observational research on eyewitness identification have identified unreliability and suggestibility in erroneous identification, and these findings are influencing the conduct of criminal investigations. A 2012 NRC report has identified ways in which new insights into individual development during adolescence may inform improvements in the juvenile justice system. At the same time, social scientific analysis has identified some areas in which it cannot provide definitive findings; for example, with respect to the deterrent effect of capital punishment.

Safer car travel

The goal of human factors engineering is developing technologies in a way that keeps humans at the center of the design process, “making the human-technology marriage work,” as human factors engineer John Lee, University of Wisconsin, Madison, described his field. Lee’s work focuses on finding ways to help people navigate a technological environment many enter every day: their cars. He studies driver distraction, an increasingly serious problem on U.S. roadways. In 2010, for example, 18% of injury-causing crashes, 3,092 deaths, and 416,000 injuries in the United States were attributed to distracted driving.

The difficulty of addressing this problem became evident in one of Lee’s studies, which tested a device designed to make it easier and less distracting to select songs on an MP3 player while driving. The study found that, contrary to expectations, drivers glanced away from the road even longer when selecting songs using the MP3 controller designed to reduce distraction than with the MP3 player alone; and the longer a driver glances away from the road, the more dangerous the situation becomes. “Engineers don’t always get it right,” Lee said.

How can the driver distraction problem be solved? One possible solution is attentive cars, which have sensors to spot obstacles in the environment and warn drivers about them, said Lee. Another option may be sensors that detect when drivers look away and cue them to get their eyes back on the road: “providing feedback so that drivers don’t have to die in order to realize that texting while driving is dangerous,” said Lee. By giving drivers reminders in the moment, he noted, attentive technology can complement the limits of regulations and education.

Human systems engineering has wide-ranging applications. Another context in which people may increasingly need help navigating technological interfaces is education, especially given the sudden emergence and rapid growth of massive open online courses (MOOCs). These courses, taught by professors at MIT, Stanford, and many other prestigious universities, are available free to any student anywhere in the world with Internet access, and they are radically extending the accessibility of higher education. What software interfaces are most useful for these courses? How should materials be presented in different areas of study for different populations? And how can valid methods of assessment be designed and implemented on a large scale with reliable security, given that MOOCs are increasingly taken for credit? Answering questions such as these will depend on findings and insights from the social and behavioral sciences.

The rest of the iceberg

These few though significant examples of the successful use of social and behavioral science to guide policies that work could easily be extended, reaching well into the multiple hundreds if we incorporate examples stretching back across more than a century: standardized testing that solved military recruitment challenges in World War I; National Accounts and sample surveys that helped to meet challenges facing the nation during the Great Depression; content analysis and area studies that contributed to national security in World War II and the Cold War; theoretical understanding of the demographic transition that was used to design global family planning that slowed population growth in the 1950s; large social experiments that revealed unintended consequences and weaknesses in social programs being implemented in the 1960s; and major research findings in early childhood learning that led to innovative parenting and schooling practices.

Despite these successes and opportunities, in one important regard the social and behavioral sciences are offering less guidance to the government and society than they could. There is not a robust understanding of the conditions that lead to the effective use of all of the sciences in the policy process: the physical and biological sciences and engineering as well as the behavioral and social sciences. Creating this understanding is a unique responsibility of the behavioral and social sciences. As documented in a recent NRC report, the use of science is itself a social phenomenon, one that should become a focus of scientific attention. Although researchers and policymakers have given much attention to randomized field trials, evidence-based-policy, and translational research, understanding of the uses of science is still too anecdotal. It is within the reach of political science, decision theory, cognitive psychology, behavioral economics, and systems theory to focus analytically on the “use of science as evidence in public policy.” This will require a much more ambitious program of research than has yet been mounted. Thus, improvement in the dissemination, communication, and use of scientific evidence looms large among the many challenges and opportunities that face the social and behavioral sciences.

“The imperative of modern society demands more attention to the social and behavioral sciences,” said National Academy of Sciences President Ralph Cicerone at the September conference. “The ambition is there, the methods are getting better, the successes are there, but not many people are hearing about it.”

Forum – Spring 2013

Reaching the middle class

In “Pathways to the Middle Class: Balancing Personal and Public Responsibilities” (Issues, Winter 2013), Isabel Sawhill, Scott Winship, and Kerry Searle Grannis present achievable early milestones sufficient for people to reach the middle class by middle age. They provide an instructive framework for understanding the circumstances that lead children and young adults to succeed—or fall down—while climbing the economic ladder. Understanding these opportunities and pitfalls helps policymakers better design and target interventions that increase children’s likelihood of success.

Consistent with academic research, Sawhill et al. emphasize the importance of early childhood success for future outcomes. Differences in school readiness by income and race appear among children as young as age 4 or 5, and early interventions can help equalize the playing field for disadvantaged children. Early childhood education programs to increase school readiness, such as Head Start, increase the chance that a child will attain a middle-class income by middle age. The evidence suggests that expanding pre-kindergarten programs more broadly would improve adult outcomes.

Sawhill et al. also emphasize the importance of family in successful childrearing. Some express concern that increased government intervention in early childhood education supplants parents’ role in raising their young children. However, the evidence from the Head Start Impact Study, in which certain children were randomly chosen to be eligible for Head Start, shows the opposite: Head Start enrollment increases the time parents spend with their children. For example, Head Start enrollment leads parents to spend 20% more time reading to their children on average, and it leads absent fathers to spend an average of around one day a month more with their children. These increases in parental involvement are sustained, lasting at least several years after children are enrolled in Head Start.

College attainment represents a later milestone in the pathway to the middle class. However, as Sawhill et al. point out, children’s college completion rates are highly correlated with parents’ socio economic status, in part because enrollment and graduation require significant family resources. This socioeconomic gap has widened over time, hardening income persistence across generations. The federal government plays a key role in increasing college attainment by expanding college accessibility through its variety of financial aid programs. Need-based financial aid, such as subsidized Stafford loans and Pell grants, specifically targets low-income students who might otherwise not be able to afford college.

As Sawhill et al. argue, success begets success and public policy interventions can prevent initial disadvantages from becoming entrenched over time. James Heckman’s work has shown that successful interventions at one stage of life increase the effectiveness of interventions at other stages. This suggests an “all-of-the-above” strategy to broadly target achieving the milestones at each stage of life identified by Sawhill et al.

JANICE EBERLY

Assistant Secretary for Economic Policy and Chief Economist
U.S. Treasury
Washington, DC

The detailed, data-packed article by Sawhill et al. is important on many fronts. Among other things, it points out that even though early intervenis especially powerful, and even though early disadvantage casts a long shadow over lifelong prospects, it is never too late to make a difference in a person’s life trajectory. It also offers numerical estimates of risk over a lifetime, which helps to answer such perennial questions as “how great is the disadvantage imposed by limited education, gender and other life circumstances? “ In the aggregate, such data help explain the concept of social mobility in America, including what is at stake and when to intervene.

One aspect of the presentation deserves particular praise, and that is the authors’ clear-eyed focus on beginning at the beginning, which is not early childhood—not “zero to three”—but rather the circumstances under which pregnancy occurs. With rare exceptions, this article being one the social interventions typically highlighted to help build stronger children and families begin after pregnancy is confirmed. For example, child and family advocates often argue for more programs to encourage early participation in prenatal care, more programs to support adequate maternal nutrition, more home-visiting programs to begin instruction on parenting, and so on. These are good ideas, on balance, but rarely do these same advocates talk openly about the circumstances surrounding conception itself; that is, when (before high school or more is completed)? with whom (a committed partner)? and under what circumstances (while in heavy debt or facing other serious challenges)? In contrast, the Sawhill et al. article makes abundantly clear that a child’s risk of becoming or remaining poor is deeply affected by maternal age and education as well as the extent to which a pregnancy is marital or nonmarital, intended or unintended—all of which are circumstances that can be ameliorated to a greater or lesser extent before pregnancy occurs.

To begin at the real beginning means that social policy would give a great deal more attention to pregnancy planning and prevention: to helping young people complete an adequate amount of education before becoming parents, to encouraging them to choose mates carefully given that raising a child is expensive and requires many years of dedication, and to avoiding unplanned, unintended pregnancy. Without a willingness to begin at the real beginning, reducing income inequality and strengthening the overall prospects of all Americans will remain less effective and more expensive than necessary.

SARAH BROWN

Chief Executive Officer
National Campaign to Prevent Teen and Unplanned Pregnancy
Washington, DC

Talk of inequality has been in political campaigns and in the media in recent months. Sawhill and her colleagues make a crucial contribution by documenting the challenges that low-income children face in succeeding in school and life: Sadly, in today’s America, economically disadvantaged children have a much lower chance of reaching the middle class than their parents and grandparents did.

With this powerful finding in mind, what becomes most important is understanding the challenges that these young people face and the opportunities that can improve their academic and vocational chances. When they experience misfortunes or inevitably make mistakes, as everyone does, the consequences can be far greater than for their wealthier peers. If they become pregnant, have a run-in with the law, or lose a part-time job, they and their parents are much less likely to have the resources to cushion the blow and smooth the road. But just avoiding trouble does not put a young person on a path to the middle class. That’s why policymakers need to focus on creating the conditions that give every child a real opportunity to succeed.

As Sawhill et al. suggest with their thoughtful recommendations for action, we know that good schools are crucial but not sufficient on their own for the long-term academic, vocational, and economic success of our nation’s young people. Research, community wisdom, and common sense tell us that families, the broader community, and schools need to work together to create integrated systems of support for each child. These supports are what America’s Promise Alliance calls the Five Promises: caring adults, safe places, a healthy start, an effective education, and opportunities to serve. Numerous empirical studies show that when young people experience these key developmental supports throughout the first two decades of life, they are more likely to exhibit good socioemotional skills, give back to their communities, and achieve academically and vocationally.

Parents, community leaders, educators, youth workers, and policymakers have focused intense effort to provide these assets. And all of these efforts are starting to pay off. According to the recently released Building a Grad Nation, the high-school graduation rate in 2010 (the most recent data available) was 78.2%, a 6.5 percentage point increase since 2001. This puts the nation on track to achieving the goal of America’s Promise Alliance’s Grad Nation Campaign; a 90% high-school graduation rate by 2020. Although graduating from high school is not sufficient to reach the middle class, it is a crucial starting point.

Americans pride themselves on living in the land of opportunity. By providing each young person with the conditions for success, we can help make the American Dream, including educational, economic, and social mobility, a reality for more and more young people.

JOHN GOMPERTS

President and CEO.
America’s Promise Alliance
Washington, DC

A new research model

In their excellent paper “RIP: The Basic/ Applied Research Dichotomy” (Issues, Winter 2013) authors Venkatesh Narayanamurti, Tolu Odumosu, and Lee Vinse take on the much-debated practice of labeling research, particularly federally supported research, as basic or applied and argue that we should put an end to such one-dimensional categorizations. They comment that the two-dimensional model of the late Don Stokes moved the discussion in the right direction but that it does not go far enough. Arguing that the innovation process is much more complex than either the basic/applied dichotomy or Stokes’ quadrants approach implies, they offer a new characterization of research as “discovery” and “invention.” They note that whatever the initial motivation, research can lead to the discovery of new knowledge about the world as well as the invention of new technologies or processes that have useful applications; indeed, both can result, and they can occur in either order or simultaneously. The authors underpin their arguments with examples in the domain of information and communication (IC) by showing how IC-relevant research discoveries and inventions that led to Nobel Prizes intersected to create new knowledge and technologies and, thus, argue that discovery and invention are not orthogonal and that the arrow between the two may reverse over time

The authors’ criticisms of the linear basic and applied model are sound, as are their arguments for seeking a more accurate model. But I would urge some caution: rapid change in long-standing federal policy often ends badly, and complicated arguments tend to be lost or misinterpreted in our political system. The suggestion of discovery and invention to describe research is an “outcomes” approach, in contrast to basic and applied, which reflect the motivation of the researcher (and often the funding entity). Moreover, as the authors point out, it can take many decades to realize research outcomes. Given this time frame, it is difficult to see how to set priorities in the allocation of funds to different fields of science or, at the grant level, how to select one proposal over another. On the matter of deciding what kind of research is appropriate for the federal government to support and what is best left to private industry, the authors suggest that “federal support is most appropriate for research that focuses on long-term projects with clear public utility,” and they note that “such research could have its near-term focus on either new knowledge or new technology.” But, while “clear public utility” might be appropriate for research in health, energy, security, the environment, and other areas of national interest, what about fundamental studies in most fields of the natural and social sciences, which industry will not support? Even in the most esoteric experimental research fields, such as observational astronomy or elementary particle physics, the interplay between fundamental studies and new technologies, many of practical importance, is richly nonlinear and unpredictable. Furthermore, I would argue that discovering new knowledge about the natural world is in the public interest.

All that said, the authors have made an important contribution to the policy discussion, and we need all the good ideas we can get. The invention and discovery approach points to the effectiveness of a particular kind of research culture, perhaps not unlike Bell Labs in the early days, where scientists and engineers in many fields worked in close proximity, sharing ideas, discoveries, and invention, and were given sufficient time to realize the fruits of their efforts. Although it is perhaps easier to create such an environment in industry than in a university or even national laboratory, the federal research funding agencies should consider how they might include such a model in their program planning, try some experiments, and assess some outcomes.

NEAL LANE

Senior Fellow in Science and Technology Policy

James A. Baker III Institute for Public Policy
Malcolm Gillis University Professor
Rice University
Houston, Texas

Better energy innovation

Michael Levi’s call for “modest” federal programs to stimulate energy innovation (“The Hidden Risks of Energy Innovation,” Issues, Winter 2013) wisely cautions against all-too-common overreaching on the part of advocates and against large, all-in programs that pick winners and losers a priori.

Given the magnitude and extent of our dependence on fossil fuels, it seems clear that there is no single-sector panacea; rather, it is more likely that a diversified portfolio of energy solutions, all sectors of which are ripe for innovation, will emerge for the future; and, given budget realities, this is consistent with the notion of modest programs. In the case of renewable energy technologies, the regional nature of renewable resources suggests that this diversified portfolio will be regional in nature, which also argues for investment in a variety of selective ways that offer high potential for payoffs.

One such regional resource that is already the focus of the sort of modest, carefully targeted federal program that Levi suggests—a resource that has received so little attention that it still flies under the radar of most analysts— is marine renewable energy: the potential of waves, tides, currents, and the oceans’ thermal structure. By providing support for research on critical topics and funding (for which significant recipient matches are required) to help bridge the “valley of death” technology gap, the U.S. Department of Energy’s Water Power Program is helping to advance the development of this underappreciated renewable resource toward becoming a full-fledged part of a future energy portfolio.

However, one unexpected, and unfortunate, challenge faced by modest programs of this nature is their invisibility relative to other, more high-profile innovation programs. The federal bureaucracy continues to be quite adept at creating significant roadblocks to progress in the form of both intraand interagency turf squabbles and foot-dragging, and the very nature of modest programs, especially their limited funding for relatively short time frames, makes them fatally vulnerable to such roadblocks. Precious dollars are spent on consultants and lobbying rather than on innovation; even more precious time is spent attempting to navigate the roadblocks.

Thus, successful implementation of Levi’s suggestions will need to include something of the nature of a top-down approach from a very high level or expeditious coordination both within and across agencies to prevent these roadblocks. The former —the “czar” approach— has obvious issues, and the latter presents an ongoing challenge for the entire Executive, one that has proven remarkably intractable in the past and is not being met now. But if modest programs are to succeed, we must do better.

HOWARD P. HANSON

Chief Scientist, Southeast National Marine Renewable Energy Center

Professor of Geosciences, Florida Atlantic University
Boca Raton, Florida

Michael Levi offers a compendium of reasons why it would be prudent of us to proceed cautiously with the proposition that public investments in technology might obviate the need for regulatory or pricing policies to address global climate change. Public investment strategies face significant ideological hurdles and risk high-profile failures. There is limited public capacity to subsidize higher-cost low-carbon technologies. And some regulatory action will likely ultimately be necessary to drive a full transition from high-carbon technologies to low-carbon technologies.

All of these are useful cautions as policymakers and advocates take stock of their options in the post–cap-and-trade world domestically and the post-Kyoto world internationally. No climate strategy will be free of political and ideological conflict. As long as low-carbon technologies cost substantially more than high-carbon technologies, neither direct subsidies nor carbon taxes and regulations are likely to be sustainable politically or economically. And even with much better and cheaper low-carbon alternatives, significant regulatory action will probably be necessary to achieve a widespread or rapid transition to a low-carbon energy economy.

The debate between the regulators and the innovators is not about whether any regulation or innovation is necessary. It is about which is the tail and which is the dog. And although most technology advocates, including the two of us, recognize the need for regulatory actions, we typically believe that lacking much better technological options, the effects of regulatory action will be modest at best and counterproductive at worst.

Whether one thinks that we have most or all of the technologies we need to mitigate climate change or are likely to develop those technologies in response to regulatory mandates turns out to be rather consequential to what one thinks we should do right now. Hence, the U.S. environmental movement, with the blessings of much of the climate and energy policy community, chose to invest somewhere north of a billion dollars over the past decade in passing a cap–and-trade program that, had it passed, would have established mandatory caps on emissions that under the best of circumstances would have had extremely modest effects on emissions and in retrospect would have established no constraint whatsoever on emissions. Thanks to the recession and the gas revolution (made possible in no small part by a three-decade public technology strategy), U.S. emissions are well below those that would have been mandated by the cap.

Levi would have us believe that the fatal flaw in this strategy was the “maximizing” conceit that passing a cap on emissions would solve both our climate and economic ills. This has become a popular trope among defenders of the environmental movement’s ill-conceived cap-and-trade effort. Had advocates, the story goes, kept the message focused on the climate issue, rather than getting distracted by the public’s demonstrable concern about the state of the U.S. economy, the effort would have fared better.

In this, Levi and others who have advanced this argument confuse messaging that was a symptom of cap-and-trade’s political toxicity with its cause. The more obvious source of cap–andtrade’s demise was the attempt to raise energy prices in the midst of the worst recession in half a century. Claims that the passage of cap-and-trade would create jobs and economic growth were a desperate attempt to turn a sow’s ear into a silk purse. The U.S. public quite rightly refused to believe that policies explicitly designed to raise energy costs would result in jobs and economic growth.

Moreover, chalking up cap–andtrade’s failure to the particular economic circumstances of the Great Recession fails to reckon with Americans’ abiding distaste for policies designed to raise their energy bills, particularly when those efforts aim to change their behavior. Cap-and-trade has now failed four times in the U.S. Congress, in good economic times and bad, and under both Democratic and Republican control. Efforts to raise Americans’ energy prices in the name of environmental protection have now been directly implicated in Democrats’ losing control of congressional majorities not once but twice over the past 20 years.

Levi suggests that the backlash against Solyndra and other failed technology investments demonstrates that technology- and innovation-centered approaches to climate mitigation are no less toxic politically. Yet despite concerted efforts by Republican operatives to turn Solyndra into a political scandal and campaign issue, not a single Democrat appears to have lost their seat over that controversy. Indeed, in contrast to cap–and-trade, which Democrats banished entirely from their vocabulary after the debacle of the 2010 elections, the President and congressional Democrats continued to tout their commitment to public investments in clean energy technology throughout the recent campaign.

Any lingering question about the relative toxicity of public technology investment policies as compared to carbon caps and taxes should have been put to rest in the weeks after the election, when the President, in his first post-election news conference, again in his inaugural address, and yet again in his State of the Union address, reiterated his intention to continue to press for energy technology investments while ruling out any renewed effort to establish carbon caps or a carbon tax as a priority in his second administration.

Levi is right to note that more highprofile failures like Solyndra risk further inflaming public skepticism about the efficacy of public technology investments. But such skepticism seems to mostly emanate from policy elites like Levi, not the public at large. Public support for government investments in energy technology, clean and conventional, has remained strong despite Solyndra and indeed through myriad failures in decades past.

What the U.S. public seems to understand, and too many energy policy analysts seem to forget, is that any successful technology innovation effort, be it public or private, must have a high failure rate if it is to succeed. Energy technology innovation is hard work and it is expensive, but when it succeeds the payoffs are huge. Just the past few years of the shale gas boom, a product of just such a costly process of trial and error (and one that was savaged in its day by soft-energy environmentalists and neoclassically trained analysts alike), will more than cover the cost of all public subsidies for all energy technologies, clean and dirty, since 1950.

Nonetheless, Levi is right that in these times of ideological polarization and fiscal austerity, energy technology advocates would be well served to look for ways to make public policies designed to accelerate innovation and support emerging technologies more effective and less costly. And they should avoid overpromising. New energy technologies take decades to develop and longer still to displace incumbent technologies. The real benefits from energy technology investments come not from creating jobs or installing above-cost, heavily subsidized technologies, but from technological revolutions that make energy cheaper and cleaner. Although the millions of jobs promised through both the green stimulus investments and cap-and-trade never materialized, the shale gas revolution is projected to create a million new jobs in the oil and gas industry by 2015 and has already generated hundreds of billions of dollars in additional economic growth in the United States annually since 2007.

In the end, despite a number of faulty assumptions, Levi does reach the right conclusion. He writes: “Indeed, a core goal of domestic technology policy should be to make regulation and other incentives easier. Anything that helps cut the cost of technology will reduce the economic burden of other policies when they are eventually pursued, making them more politically palatable.”

Those who believe that energy innovation is the central challenge that all climate mitigation strategies must address could not have said it better. With cheap, clean energy technology, anything is possible; without it, to paraphrase Macbeth, climate politics and policy amount to much sound and fury, signifying nothing.

TED NORDHAUS

Chairman

MICHAEL SHELLENBERGER

President
The Breakthrough Institute Oakland, CA

Energy conversion

In “Four Technologies and a Conundrum: The Glacial Pace of Energy Innovation,” (Issues, Winter 2013), Michael E. Webber, Roger D. Duncan, and Marianne Shivers Gonzalez provide useful insights into the U.S. energy system by re-casting the conventional picture of energy consumption by resource and end-use sector into one that highlights the major technologies we employ to convert primary energy into more usable forms. In this framework, they identify four technologies that together provide two-thirds of U.S. energy: steam and combustion turbines (primarily for electricity generation), and two types of internal combustion engines (spark- and compression-ignition, used mainly for transportation). All of these technologies, they note, date from the late 18th and 19th centuries (as do others on their list). Thus, they see that “fundamental innovation in power and transportation technology is slow” and call for increased R&D investments concentrated on new energy conversion pathways that reduce waste heat, along with policies that can accelerate the development, commercialization, and adoption of new technologies at scale.

Certainly there is no denying that the U.S. energy sector in general, and the electric power industry in particular, have been laggards in the pace of innovation, as reflected by such measures as patenting activity, R&D spending, and the size of the R&D workforce. Data we assembled for a recent National Academies study, for example, estimated that the ratio of R&D spending to sales was in the range of only 0.2 to 2.2% for energy-related U.S. industries, with utilities near the low end of that range. This compared to roughly 10 to 17% for industries such as software/ computer services, pharmaceuticals, and biotechnology. Similarly, the percentage of R&D scientists and engineers in the U.S. utilities labor force (0.3%) was minuscule compared to the most innovative industries (10 to 20%) and far smaller than the average for all U.S. industries (7.1%). Equally sobering has been the sharp decline in federal energy R&D spending over the past three decades, both in absolute terms and as a percentage of all federal nondefense R&D. As a result, the United States spends less on energy R&D than many other industrialized countries as a percentage of gross domestic product, and less than some smaller countries (such as Japan) on an absolute basis as well.

The call by Webber et al. for measures to reverse this trend and accelerate the pace of energy technology innovation thus rings true, and echoes other recent studies. The authors don’t say, however, exactly what policy measures besides R&D investments they would recommend, nor do they elaborate on why folks other than engineers should care about thermodynamic efficiency. Implicit in their discussion, however, is the potential for new energy-conversion technologies to dramatically reduce the rate of natural resource depletion and move us toward a zero-carbon economy based on renewable energy sources. If these in fact are the societal goals we seek, achieving them any time soon will require not only policy “carrots” (such as more R&D) but also policy “sticks” that significantly limit greenhouse gas emissions and create markets for efficient new lowcarbon technologies. Webber et al. call it “imperative” that appropriate policies be put in place soon, an urgency underscored in the America’s Climate Choices study.

Toward that end, the energy accounting scheme they illustrate provides a useful perspective that can help identify major targets of opportunity for technological change. Their approach would be even more useful if electricity generation were treated not as a final end-use sector but as the intermediate process it is. That could help identify other important devices and conversion technologies (such as motors, servers, and incandescent lamps), in which efficiency gains could further contribute to environmental goals, along with other types of innovations. Finally, we should heed the authors’ admonition to not expect new “breakthrough technologies to magically solve the problem. Although the discovery of novel, radical, breakthrough, leapfrog, out-of-the-box, game-changing technologies is what all energy R&D managers hope for, what in retrospect is seen to be a major innovation is more often than not the product of many years of incremental improvements that could not have been achieved overnight. Together with support for new discoveries and inventions, the sustained pursuit of such improvements should be the bedrock of U.S. policies to foster long-term energy innovations.

EDWARD S. RUBIN

Alumni Chair Professor of Environmental Engineering and Science

Professor of Engineering and Public Policy and Mechanical Engineering

Carnegie Mellon University Pittsburgh, Pennsylvaniarubin

Keeping kids in school

“Staying in School: A Proposal for Raising High-School Graduation Rates” (Issues, Winter 2013) by Derek Messacar and Phillip Oreopoulus hits on an issue that every state and chief state school officer has on the education reform agenda. In Kentucky, we have had legislation proposed to raise the dropout age from 16 to 18 for the past four years. Governor Steve Beshear and First Lady Jane Beshear have been tremendous champions of this legislation, and we have typically been able to move the legislation through the House of Representatives. However, the legislation has stalled in the Senate. This year, however, there are positive signs for a compromise on the legislation.

As a former school superintendent whose district was very successful in lowering the dropout rate, I was at first confused by the resistance of legislators and educators in Kentucky to raising the dropout age. The resistance centered on several key issues. First and foremost was the concern about an unfunded mandate created by raising the dropout age. Superintendents were concerned that raising the dropout age would mean that over 6000 more students (16-, 17-, and 18–year-olds) would remain in school, and the state did not have funds to support these students. The next concern was behavior. Many teachers expressed concern that forcing students to stay in school who did not want to be in school would create unsafe schools and classrooms. Also, numerous legislators suggested that the research does not support raising the dropout age. Simplistic correlation studies among states that have raised the dropout age show little to no impact on graduation rates. Finally, the major concern was the lack of funding and resources to support the needed alternative and career/technical programs.

The Messacar and Oreopoulos article provides significant research and data to address the concerns that were voiced in Kentucky and across the nation. A key issue that the article brings to the front is that raising the dropout age has a positive impact on the state’s economy and also lowers a state’s incarceration costs and social program costs. In the long run, raising the dropout age is a revenue producer and not an unfunded mandate.

The national blueprint proposed by the authors does provide concrete suggestions for states and chief state school officers as they address the dropout issue. Ensuring that every child graduates from high school with the skills needed to be successful in college and careers or the military is the moral, civil rights, safety, and economic imperative of our generation.

TERRY HOLLIDAY

Commissioner of Education
Kentucky Department of Education
Frankfort, Kentucky

We can all agree that long-term solutions to our nation’s jobs crisis require us to build a world-class workforce. To do this, we must think strategically about keeping kids in school, and we must ensure that all students graduate from high school on time, ready to enter college or start their careers. As suggested by Messacar and Oreopoulos, nothing short of a comprehensive approach will adequately address our nation’s high-school dropout problem.

Part of any multi-tiered solution should suggest that we transform the school experience to be more relevant and rigorous for students. As a former biology teacher, career and tech-ed high-school principal, and local district superintendent, I’ve seen firsthand the opportunities that exist when students are given more ways to learn in challenging and engaging classroom environments. Project Lead the Way is a terrific example of a STEM (science, technology, engineering, and mathematics) program that equips students with the skills necessary to compete for high-skill, high-wage jobs.

There’s no question that the most powerful component of any school is the effectiveness of its educators. A great teacher has a lifelong impact on his or her students—in fact, I consider this the most important factor in addressing high-school graduation and dropout rates. We can foster a culture of excellence by ensuring that every classroom has a great teacher, every school has an effective principal, and every district has dynamic local leadership. Through professional evaluation systems and transparent school performance measures, we can recognize local schools for their work.

Another opportunity to decrease dropout rates lies in our ability to find new ways to engage parents. The first step any educator, community member, or legislative leader can take is to advocate for policies that give parents the freedom to choose the school that best meets their child’s needs. Whether it’s through charter schools, magnet schools, or traditional public school transfers, school choice provides an important opportunity to meaningfully engage parents.

Finally, as a state education chief, I understand that it’s tempting to become prescriptive in any well-intentioned effort to decrease high-school dropout rates. While there is no silver bullet that will solve this problem, if we hold ourselves accountable for the results we hope to see in education, and if we give local schools the flexibility to compete, then we will spur innovation at the district level. Messacar and Oreopoulos make a fine argument in favor of increasing the mandatory attendance age, but I fear that unless we transform the school experience in a comprehensive manner, we will continue to elicit marginal results.

TONY BENNETT

Commissioner
Florida Department of Education
Tallahassee, Florida

I enjoyed reading “Staying in School.” I would like to share some additional points to help us continue this critical dialogue about the importance of high-school graduation and dropout prevention.

Many of your readers may not know that the national dropout rate the authors refer to does not include the number of students who drop out but later return to school. For this reason, nearly every state in the nation now calculates and reports four- and five-year cohort graduation rates. These rates reflect the percentage of students who enter high school and graduate four or five years later. In North Carolina, our four-year cohort graduation rate hit an all-time high of 80.4% in 2011–2012. This milestone also represents an increase of 12.1 percentage points since the state first reported a four-year cohort rate of 68.3% in 2006.

In addition, the United States saw a 78.2% four-year graduation rate in 2010, which is up 6.5 percentage points from the 71.7% national rate in 2001, according to the 2013 Building a Grad Nation report released last month. Although there is certainly more work to be done, these achievements at the state and national levels are important to note.

The significant improvement in North Carolina’s graduation rate certainly did not happen by chance. We have implemented many initiatives aimed at keeping students in school, and our hard work is paying off. For example, we have offered many students the opportunity to earn industry-recognized career certifications through our Career and Technical Education program. These programs help students make a real connection between what they are learning in the classroom today and the jobs they will be applying for in the future. Students earned more than 91,000 of these certifications last year, and we are hoping for an even higher number in 2013.

We also are providing many public high-school students an opportunity to get a head start on a college education, for free. North Carolina currently leads the nation with its number of early college high schools, where students can earn a high-school diploma and up to two years of college credit in just four years.

And most recently, we have revised and updated our state’s entire curriculum to reflect the knowledge and skills students must master at each grade level, so they can be successful in higher education and the workplace. We have done all of this work to move us closer to reaching our ultimate goal of a 100% high-school graduation rate.

I appreciate your highlighting the issue of raising the compulsory attendance age. This is certainly one method among many that could help keep more students in school. And while we consider this strategy, let’s also remember to recognize the progress we have made and the successes we have seen so far, so we can ensure not only that more students earn diplomas, but also that they are leaving high school prepared for college, careers, and citizenship.

JUNE ST. CLAIR ATKINSON

State Superintendent
Public Schools of North Carolina
Raleigh, North Carolina

“Staying in School” aptly identifies the complexities as well as the policy opportunities surrounding state and federal efforts to increase high-school graduation rates. I launched Graduation Matters Montana in 2010 to dramatically decrease the number of Montana high-school students who drop out each year. Through a private/public partnership, 28 communities have launched a Graduation Matters initiative, meaning that local schools, community organizations, and main street businesses are working together to address their community’s dropout rate and to develop more college and career pathways.

Early signs are that the strategy is working: 65% of Montana high-school students now attend a Graduation Matters school, and our state dropout rate has gone from 5.1 to 4.1%.

As state superintendent, I am pursuing all angles, from changing state policies to equipping local community efforts through Tool Kits, technical assistance, and privately funded incentive grants. Our state still allows students to drop out at age 16, and I believe we can do better; yet our state legislature has twice rejected proposals to change that state law. I will not wait for state or federal lawmakers to act. Graduation Matters Montana is equipping local community efforts to support student success.

For more information on our community-based strategy, go to http://graduationmatters.mt.gov/, which contains many resources that are particularly pertinent for rural states and rural communities.

DENISE JUNEAU

Montana Superintendent of Public Instruction
Helena, Montana

From the Hill – Spring 2013

Sequestration follies

In spite of substantial debate and controversy in the days leading up to the deadline, sequestration went into effect on March 1 as required by law. Cuts to defense and nondefense R&D will total an estimated $9.0 billion in the FY 2013 federal R&D budget. These cuts will be particularly harsh because they must be squeezed into the final seven months of the fiscal year.

Some agencies have already issued new memoranda on the impact of sequestration on agency operations. The National Institutes of Health (NIH), for example, stated that the “impact could include: not issuing continuation awards, or negotiating a reduction in the scope of…awards to meet the constraints imposed by sequestration. Additionally, plans for new grants or cooperative agreements may be rescoped, delayed, or canceled depending on the nature of the work and the availability of resources.” The National Science Foundation (NSF), meanwhile, has stated that although it will honor existing grants, “the total number of new research grants will be reduced by approximately 1,000.”

As widely expected, Congress failed to pass a more balanced alternative deficit reduction plan to replace the sequester. The House of Representatives, however, did pass an appropriations bill (H.R. 933) that would fund the federal government for the remainder of the 2013 fiscal year. The legislation, which passed 267-151, would fund both defense and nondefense R&D at FY 2012 levels, but because the sequester remains in effect, the net effect is a roughly 7.8% decrease for defense R&D and a 5% decrease for nondefense agency R&D. The House bill was written as an appropriations bill for the Departments of Defense (DOD) and of Veterans Affairs (VA), but as a continuing resolution (CR) for the remaining agencies. This allowed the House to provide some flexibility to DOD and VA on how each could allocate the sequester cuts.

On March 11, the Senate Appropriations Committee released its revised version of the House bill. Both Appropriations Chairwoman Barbara Mikulski (D-MD) and Ranking Member Richard Shelby (R-AL) agreed to the legislation that would continue to fund the government for the remainder of the fiscal year. The revised “hybrid” bill does not eliminate the sequestration but does expand on the House version by including additional flexibility for the Departments of Justice, Homeland Security, Agriculture, and Commerce, as well as the National Aeronautics and Space Administration and NSF. The bill also includes a small increase ($71 million, pre-sequestration cut) for the National Institutes of Health. On March 20, the Senate voted 73-26 to pass its version of the Continuing Appropriations Act. The Act includes an amendment submitted by Senator Tom Coburn (ROK) that limits funding for political science research at the NSF; specifically, the agency will be able to fund political science research only if it is certified by the NSF director as “promoting national security or the economic interests of the United States.” The next day, the House of Representatives voted 318-109 to approve the Senate’s changes.

Congress In Brief

On Jan. 28, a bipartisan group of eight senators released a set of principles for immigration reform at a press conference. The draft framework includes a proposal to “award a green card to immigrants who have received a PhD or master’s degree in science, technology, engineering, or math from an American university.” It further states that “It makes no sense to educate the world’s future innovators and entrepreneurs only to ultimately force them to leave our country at the moment they are most able to contribute to our economy.” In addition, it will establish an easier path to citizenship for illegal immigrants who came to the United States as minors—similar to the DREAM Act. This final point is expected to be a point of contention as the debate unfolds, especially as the draft principles still need to be crafted into legislation (more background found here). The bipartisan “Gang of Eight” who crafted the principles are Senators Charles Schumer (D-NY), John McCain (R-AZ), Lindsey Graham (R-SC), Michael Bennet (D-CO), Marco Rubio (R-FL), Dick Durbin (DIL), Robert Menendez (D-NJ), and Jeff Flake (R-AZ). Obama subsequently lauded the Senate plan and released his own principles on immigration reform.

The Senate Budget Committee has established a website for citizen input called MyBudget, which provides Americans the opportunity to share their stories, identify their budget priorities, and provide ideas for how to achieve fiscal reform. Participating on the site may be of particular interest to those in the science and innovation community who have firsthand experience with federal R&D funding.

Sen. Jay Rockefeller (D-WV) has introduced S.21, the Cybersecurity and American Cyber Competitiveness Act of 2013. Sponsors of the bill hope “to secure the United States against cyber attack, to improve communication and collaboration between the private sector and federal government, to enhance American competitiveness and create jobs in the information technology industry, and to protect the identities and sensitive information of American citizens and businesses.” The bill also includes language that promotes R&D investments to expand the IT workforce and improve the U.S. economy.

The Senate Health, Education, Labor, and Pensions Committee held a hearing on January 24 to assess the state of the U.S. mental health system. Six witnesses testified, including Tom Insel, head of the National Institute of Mental Health, who noted that about one in five Americans is affected by mental illness.

Congressional committee changes

The beginning of any new Congress can feel like a game of musical chairs, and the 113th is no exception. What follows are some notable shifts in committee and subcommittee chairs.

Sen. Barbara Mikulski (D-MD) will chair the Senate Appropriations Committee. Mikulski has been the longtime head of the Commerce-Justice-Science appropriations subcommittee, which funds a number of key federal research agencies—for example, the National Science Foundation. Sen. Thad Cochran (R-MS) is the ranking member. On the House side, Hal Rogers (R-KY) returns to chair the House Appropriations Committee and Nita Lowey (D-NY) will serve as the new ranking member. The committee recently announced its majority and minority subcommittee leaders. Rep. Jack Kingston (R-GA) takes over for former Rep. Denny Rehberg as chairman of the Labor-HHS Subcommitee, which funds the National Institutes of Health (NIH).

Rep. Lamar Smith (R-TX) will chair the House Science, Space and Technology Committee. Smith is former head of the Judiciary Committee and was heavily involved in congressional patent reform efforts. Continuing on as ranking member will be Rep. Eddie Bernice Johnson (D-TX). New subcommittee chairs are listed here (majority) and here (minority). In the Senate, Jay Rockefeller (D-WV) returns to the helm of the Commerce, Science and Transportation Committee, while John Thune (R-SD) takes over as ranking member. Rockefeller recently announced that he plans to retire in 2014.

Sen. Tom Harkin (D-IA) returns to his spot atop the Senate Health, Education, Labor and Pensions Committee, while Lamar Alexander (R-TN) has signed on as ranking member. Like Rockefeller, Harkin has announced that this will be his last term in Congress. Harkin is widely known as a strong supporter of federal biomedical research.

In other news, some new caucuses have formed on Capitol Hill. On Dec. 7 Rep. Randy Hultgren (R-IL) announced the creation of the bipartisan House Science and National Labs Caucus. The caucus’s focus is to raise awareness about the role that federal labs play in long-term economic growth. Other co-chairs include Reps. Chaka Fattah (D-PA), Ben Ray Lujan (D-NM), and Alan Nunnelee (R-MS). A few days later, Senate Environment and Public Works Committee Chair Barbara Boxer (D-CA) announced that she would form a climate change caucus in the wake of Hurricane Sandy. Boxer said of the new caucus, “it is going to work with all the committees and all the committee chairmen to make sure we can move forward legislation that reduces carbon pollution and also works on mitigation and all of the other elements… I think you are going to see a lot of bills on climate change.”

“From the Hill” is adapted from the newsletter Science and Technology in Congress, published by the Office of Government Relations of the American Association for the Advancement of Science (www.aaas.org) in Washington, DC.

The Path to Continuously Learning Health Care

Health care in the United States has experienced an explosion in biomedical knowledge, dramatic innovations in therapies and surgical procedures, and expanded capacity to manage conditions that previously were debilitating or fatal—and ever more exciting H clinical capabilities are on the horizon. Yet, paradoxically, health care is falling short on basic dimensions of quality, outcomes, cost, and equity. Actions that could improve the health care system’s performance—developing knowledge, organizing and translating new information into medical evidence, applying the new evidence to patient care—are marred by significant shortcomings and inefficiencies that result in missed opportunities, waste, and harm to patients.

The human and economic impacts are great. An estimated 75,000 deaths could have been averted in 2005 alone if every state had delivered care on par with the best performing state. Current waste—an estimated $750 billion in unnecessary health spending in 2009—diverts valuable and limited resources from productive use.

It is important to note that individual physicians, nurses, technicians, pharmacists, and others involved in patient care work diligently to provide high-quality, compassionate care to their patients. The problem is not that they are not working hard enough. Rather, it is that the health care system does not adequately support them in their work. The system lags in adjusting to new discoveries, disseminating data in real time, organizing and coordinating the enormous volume of research and recommendations, and providing incentives for choosing the smartest route to health, not just the newest, shiniest—and often most expensive—tool. These broader issues prevent clinicians from providing the best care to their patients and limit their ability to continuously learn and improve.

The shortcomings are especially apparent when considering how other industries routinely operate compared with many aspects of health care. Builders rely on blueprints to coordinate the work of carpenters, electricians, and plumbers. Banks offer customers financial records that are updated in real time. Automobile manufacturers produce thousands of vehicles that are standardized at their core, while tailored at the margins. Although health care may face unique challenges in accommodating many competing priorities and human factors, the health care system could learn from these other industries how to better meet specific needs, expand choices, and shave costs.

The bottom line is that the nation, and its citizens, would be better served by a more nimble health care system that is consistently reliable and that constantly, systematically, and seamlessly improves. In short, the nation needs health care that learns by avoiding past mistakes and adopting newfound successes.

A vision—and a pathway

The Institute of Medicine has provided a roadmap for reaching this goal. It is detailed in Best Care at Lower Cost: The Path to Continuously Learning Health Care in America, a report released in September 2012. The good news is that opportunities for improving health care exist that were not available just a decade ago. Vast computational power is increasingly affordable and widely available, and connectivity allows information to be accessed in real time virtually anywhere by professionals and patients, permitting unprecedented diffusion of information cheaply, quickly, and on demand. Human and organizational capabilities offer expanded ways to improve the reliability and efficiency of health care. And health care organizations and providers increasingly recognize that effective care must be delivered by collaborative teams of clinicians, each member playing a vital role.

Yet simply acknowledging such opportunities does not necessarily result in putting them to good use. Indeed, building a learning health care system within current clinical environments requires overcoming substantial challenges. Clinicians routinely report moderate or high levels of stress, feel there is not enough time to meet their patients’ needs, and find their work environments chaotic. They struggle to deliver care while confronting inefficient workflows, administrative burdens, and uncoordinated systems, preventing them from focusing on additional tasks and initiatives, even those that have important goals for improving care.

Given such real-world impediments, crafting and implementing initiatives that focus merely on incremental improvements and add to a clinician’s daily workload are unlikely to succeed in fundamentally improving health care. Significant change can occur only if the environment, context, and systems in which health care professionals practice are reconfigured to support learning and improvement.

Realizing these objectives will require efforts in four main areas: generating and using real-time knowledge to improve outcomes; engaging patients, families, and communities; achieving high-value care; and creating a new culture of care.

Advancing real-time knowledge

Although unprecedented levels of information are available, clinicians and patients often lack practical access to guidance that is relevant, timely, and useful for the circumstances at hand. For example, of the clinical guidelines for the nine most common chronic conditions, fewer than half address the issues of patients who experience two or more of the conditions at the same time, even though 75 million patients fit this category. Bridging gaps in how knowledge is gathered and used will require applying computing capabilities and analytic approaches to develop real-time insights from routine patient care and then using new technological tools to disseminate the emerging knowledge.

One key step will be to strengthen the digital infrastructure of the health care system to better capture data on clinical care and patient outcomes, on the care delivery process, and on the costs of care. Data should be digitally collected, compiled, and protected as reliable and accessible resources for managing care, assessing results, improving processes, strengthening public health, and generating new knowledge.

Large quantities of clinical data are now generated every day in the regular process of care, but most of the information remains locked inside paper records that are difficult to access, transfer, and query. Digital systems have the potential to turn each of those bothersome traits on its head. Care must be taken, however, to integrate the new electronic methods seamlessly into providers’ daily workflow so as not to disrupt the clinical routine.

To complement the development of better digital systems, efforts are needed to promote expanded access to data and expanded data sharing. The idea is that the capacity for learning experiences increases exponentially when a system can draw knowledge from multiple sources. In one promising example, called distributed data networks, each participating organization stores its information locally, often in a common format. When a researcher seeks to answer a specific research question, all of the organizations in the network execute identical computer programs that analyze the data, create a summary from each site, and share those summaries with the entire network. In other efforts to expand data collection and access, insurance companies and other payer groups, health care delivery organizations, and companies that make medical products should be encouraged to contribute data to ongoing and new research efforts. And patients can play an important role by fully participating in self-reporting systems designed to gather data on patient outcomes, and by using new communication tools, such as personal portals, to better manage and record their own care.

Beyond technical matters, various legal and regulatory restrictions can be barriers to real-time learning and improvement. The privacy and security rules under the Health Insurance Portability and Accountability Act (HIPPA) pose particular challenges. In several surveys, researchers have reported that the rules increase the time and cost of research, impede collaboration among researchers, and make it difficult to recruit volunteers for studies. Protecting patient privacy is, of course, the basic starting point. But the current rules, with their inconsistent interpretation, offer a relatively limited security advantage to patients while impeding health research and the improvement of care. HHS is currently reviewing HIPPA rules, along with the policies of various institutional review boards that oversee research at many locations, with respect to actual or perceived regulatory impediments to the use of clinical data.

As more and better data become available, the obvious job will be to identify and adopt improved approaches for delivering accurate information to clinicians and patients in a timely manner. This will require making decision support tools and knowledge management systems routine features of health care delivery. Accelerating their use requires developing tools that deliver reliable, current clinical knowledge, in a clear and understandable format, to providers at the point of care, in addition to incentives that encourage the use of these tools. This also requires a shift in health professional education to teach skills for engaging in lifelong learning on how best to deliver safe care in an interdisciplinary environment. Furthermore, there are still multiple poorly understood barriers to dissemination and use of scientific evidence at the point of care. Addressing these barriers will require additional research and the development of practical tools that can improve the usefulness and accessibility of such data for clinicians and patients.

Empowering patients

An effective, efficient, and continuously learning system requires patients who are actively engaged in their own care. Clinicians supply information and advice based on their scientific expertise in treatment and their best assessment of potential outcomes, while patients, their families, and other caregivers bring personal knowledge on the suitability—or lack thereof—of different treatments for the patient’s circumstances and preferences. Both perspectives are needed to select the right care. Of course, providing what has come to be called “patient-centered” care does not mean that providers simply agree to every patient request. Rather, it entails meaningful awareness, discussion, and engagement among patient, family, and the care team on the evidence, risks and benefits, options, and decisions in play.

The structure, incentives, and culture of the current health care system, however, are poorly aligned to engage patients and respond to their needs—and patients are often insufficiently involved in their care decisions. Even when encouraged to play a role in decisions about their care, they often lack understandable, reliable information—from evidence on the efficacy and risks of different treatment options to information on the quality of different health care providers and organizations—that is customized to their needs, preferences, and health goals.

Patient-centered care takes on increasing importance in light of research that links such care to better health outcomes, lower costs, and customers—the patients themselves— who are happier with their experience, among other benefits. With these rewards in mind, health care providers and organizations will need to draw on a full toolkit of actions.

Providers should begin by placing a higher premium on involving patients in their own health care to the extent that patients choose, encouraging them and their families to be active participants. From this base, clinicians should employ high-quality, reliable tools and skills that are customized to a patient’s situation to aid in shared decision making. New technologies offer opportunities for clinicians to engage patients by meeting them where they are, rather than in traditional clinical settings. Further efforts may include providing new online sources of information and assisting patients in managing their own health—options that highlight the need for health professionals to assume new roles in partnering with patients.

Several actions can increase patient centeredness more broadly. First, there is a need for new tools that can assist individuals in managing their health and health care. Furthermore, public and private payers can promote and measure patient-centered care through payment models, contracting policies, and public reporting programs. There are also gaps in our ability to measure patient-centered care, which will require the development of a reliable set of measures of patient-centeredness for consistent use across the health care system. These measures can be used both to incentivize patient centered care and to assist organizations as they measure their improvement.

Fostering high-value care

Health care payment policies strongly influence how care is delivered, whether new scientific knowledge and insights about best care are diffused broadly, and whether improvement initiatives succeed. The prevailing approach to paying for health care, based predominantly on paying set fees for individual services and products, encourages wasteful and ineffective care. New models of paying for care and organizing care delivery are emerging to improve quality and value. Although evidence is conflicting on which models work best and under what circumstances, it is clear that a learning health care system would incorporate incentives aligned to encourage continuous improvement, identify and reduce waste, and reward high-value care.

The system would also be transparent. It would systematically monitor the safety, quality, processes, costs, and outcomes of care and make the information available for clinicians, patients, and families to use in making informed choices. This type of information on health care options, quality, price, and outcomes can then spur conversations among individuals and health care providers to promote informed decision making.

Health care delivery organizations and clinicians should fully and effectively employ digital systems that capture patient care experiences reliably and consistently.

Multiple strategies exist for increasing the value of health care. Health care delivery organizations can use systems engineering tools and process improvement methods to eliminate inefficiencies, remove unnecessary burdens on clinicians and staff, enhance patient experience, and improve patient health outcomes. Furthermore, these organizations can reward continuous learning and improvement through internal practice incentives. For their part, public and private payers can adopt outcome- and value-oriented payment models and contracting policies that would support high-quality, team-based care focused on the needs and goals of patients and families. With an eye toward ongoing improvement, payment models, contracting policies, and benefit designs need to continuously refined to better reward high-value care that improves health outcomes.

Creating a new culture

Although each step along the path to a learning health care system is important, none by itself is sufficient. Rather, the host of needed changes must be interwoven to achieve a common goal: health care organizations that are devoted at their very core to optimizing care delivery practices, continually improving the value achieved by care and streamlining processes to provide the best patient health outcomes. Reaching this point will require broad participation by patients, families, clinicians, care leaders, and those who support their work. Health care delivery organizations, however, will play an especially important role. Because of their size and care capacities, they can set an example for improvement across the health care system by using new practice methods, setting standards, and sharing resources and information with smaller facilities and individual care providers.

Although details may vary among organizations, some key concepts will remain constant. A learning health care organization harnesses its internal wisdom—staff expertise, patient feedback, financial data, and other knowledge— to improve its operation. It also engages continuous feedback loops monitoring internal practices, assessing what can be improved, testing and adjusting it response to data, and implementing its findings across the organization.

Simply put, an organization that promotes continuous learning and improvement is one that makes the right thing easy to do. Its environment simplifies procedures and workflows so that providers can operate at peak performance to care for patients, and embraces support tools, such as checklists, that make providers’ jobs easier. This not only improves care delivery and patient outcomes; it also reduces stress on front-line care providers, improves job satisfaction, and prevents staff burnout.

Many organizations still struggle to implement such transformational system changes. They face both external obstacles, such as financial incentives that emphasize quantity of service over quality, and internal challenges to achieving constant improvement. To evolve successfully, health care organizations must develop a culture that supports improvement efforts, by adopting systematic problem-solving techniques, building operational models that encourage and reward sustained quality, and becoming transparent on costs and outcomes.

Leadership will be vital, as an organization’s leadership and governance set the tone for the entire system. The visibility of leaders at the highest level makes them uniquely positioned to define the organization’s quality goals, communicate these goals and gain acceptance from the staff, and make learning a priority. Leaders also have the ability to align activities to ensure that staff members have the necessary resources, time, and energy to accomplish the organization’s goals. By defining and visibly emphasizing a vision that rewards continuous learning and improvement, leadership encourages an organization’s disparate elements to work together toward a common end.

To complement leadership at the top, a continuously learning organization also requires leadership on the part of the managers and front-line workers who translate an expressed vision into practice. Middle managers play a crucial role in on-the-ground, day-to-day management of the units that, collectively, make up an organization. Unit leaders therefore must often challenge the prevailing mental models—deep-seated assumptions and ways of thinking about problems—and refocus attention on the barriers to learning and improvement. To this end, middle managers must be able to set priorities for improvement efforts, establish and implement continuous learning cycles, and foster a culture of respect among staff that empowers them to undertake continuous learning and improve patient care.

To promote continuous learning, health care organizations also need to adopt dedicated learning processes— mechanisms that help in constantly capturing knowledge and using the lessons to implement improvements. Achieving this type of systems-based problem solving requires an organizational culture that incentivizes experimentation among staff. While success is the goal, the system should recognize failure as key to the learning process and not penalize employees if their experiments are unsuccessful. Systems that continuously learn also need to be adept at transferring the knowledge they gain throughout the organization. Although each of these factors is important, it is the organization’s operational model—the way it aligns goals, resources, and incentives—that makes learning actionable. In this way, an organization’s operating model can promote continuous learning, help control variability and waste that do not contribute to quality care, and recoup savings to invest in improving care processes and patient health, and make improvement sustainable.

Pioneering health care organizations that successfully become continuously learning operations—fully or even partially— should also take the lead in diffusing the lessons learned more broadly. In this way, they not only can stand as beacons of opportunity, but also can provide the type of granular, hard-won information that can encourage and speed similar transformations across the entire health care system.

The entrenched challenges of the U.S. health care system demand a transformed approach. Left unchanged, health care will continue to underperform; cause unnecessary harm to patients; and strain national, state, community, and family budgets. The actions required to reverse current trends will be notable, substantial, sometime disruptive—and absolutely necessary.

The challenges are clear. But the imperatives are also clear, the changes are possible, and there are at least signs of success. Moving ahead, following the path to a continuously learning health care system, offers the prospect for best care at lower cost for individuals and society.

Forum – Winter 2013

Energy basics

Bruce Everett (“Back to Basics on Energy Policy,” Issues, Fall 2012) reviews the history of government energy policy with discomforting accuracy. One can only hope that the article will persuade more of us that Walt Disney’s first law, “Wishing will make it so,” is fine for feel-good cartoons, but a very poor guide for policy. As the article makes clear, government can take rightful pride in its support of basic research on energy-related science and early-stage engineering. But grandiose schemes to remake our energy economy have not worked.

It is hard to quibble with the main ideas of “Back to Basics,” but successful technologies don’t always follow “the three distinct stages” of basic research, technical development, and commercialization that Everett mentions. For example, James Watt had only a rudimentary understanding of basic thermodynamics as he developed the high-efficiency steam engines that accelerated the Industrial Revolution. And in many of his most important inventions, Thomas Edison was not particularly concerned with having “a solid understanding of the science involved.” Neither Watt nor Edison became bogged down by government interactions. In fact, they hardly interacted with their governments at all, except in the important area of intellectual property.

Discussing externalities, Everett mentions that “Climate change scientists argue that increasing CO2 will have a catastrophic impact on humanity.” Quite a few distinguished scientists, including climate scientists, do not agree with this apocalyptic assessment. There are even credible arguments that more CO2, with its modest warming, will be a net benefit to humankind, for example, because of increased agricultural yields. The movement to demonize CO2 has many of the trappings of religious zealotry. One wonders what future historians will make of the cult-like hectoring of our citizens to reduce their “carbon footprints” or the uncritical promotion of “renewables,” including wind, solar, and ethanol, soon to be available in abundance from cellulose, in accordance with Walt Disney’s first law and congressional legislation.

“Back to Basics” has some of the flavor of Thomas Paine’s Common Sense, published in 1776, which reviewed the British government’s colonial policies in North America. Everett’s equally commonsensical review makes a persuasive case that our government’s policy on energy is in equal need of reform.

WILLIAM HAPPER

Cyrus Fogg Brackett Professor of Physics
Princeton University
Princeton, New Jersey

The author served as the director of the Department of Energy’s Office of Energy Research (now the Office of Science) from 1990 to 1993.

I find myself in agreement with Bruce Everett that some energy policies have been misguided, but find that he misses the overall reality. I would argue that there is little evidence of an efficient energy “market” that is effectively choosing among fuels and technologies or that through some mysterious process solves our national security and environmental problems.

First, his characterization of the policies that have promoted nuclear power and biofuels are quite compelling. In fact, he even shortchanges his argument by leaving out the market failure of insurance for nuclear power and the liability limitations of the Price Anderson Act. He also underplays just how flawed biofuels policy is in the United States and Europe, where by default, burning biofuels is assumed to be as carbon-neutral as wind, solar, and geothermal. Carbon credits are provided to low-efficiency woody biomass electricity plants despite their emitting 30% more CO2 per megawatt-hour than a coal steam turbine. Massachusetts may be the only political entity to have established its criteria for biomass emission credits on a scientifically sound basis.

However, Everett argues that even well-designed and targeted policies to induce changes in the national energy mix are misguided. He argues that cost and the market alone should decide what our energy supply mix should be, and government should retreat to only supporting basic research. He correctly acknowledges that there are circumstances in which government intervention may be needed to correct market externalities such as national security and pollution, but he does not consider climate change to be in that category despite the extent and irreversibility of its impacts. He also dismisses the Pentagon’s interest in efficiency, renewables, and alternative fuels as unnecessary to the military mission of defending our country, which requires a diverse and secure fuel supply. It makes military sense to replace diesel generators in combat with solar panels and lightweight batteries rather than sacrifice troops in the protection of fuel convoys in Afghanistan. He also ignores the past century of cash subsidies and policy advantages that the fossil fuel industry enjoys.

The International Energy Agency reports that globally, fossil fuel subsidies grew to more than $400 billion in 2010. Subsidies and tax breaks for oil, gas, and coal have averaged $3 billion to $5 billion per year in the United States. In addition, the indirect subsidies in the United States for below-market drilling and mining leases on federal lands, allowing mountaintop removal for coal extraction without restoration, the current lax enforcement of environmental rules for protecting water and land from gas and oil fracking, and the unregulated pollution that is permitted under air and water laws is not internalized. Coal is cheap simply because it does not pay the full cost of the damage suffered by others in its extraction and combustion. The vulnerability of our economy to wild spikes in world oil prices has preceded most of the recessions of the past half-century. Our heavy military investment to keep sea lanes open for oil transport is paid out of general revenues and not through an energy tax. Is it not in the national interest to reduce our vulnerability to oil supply disruptions and to spend a bit to reduce oil demand by making homes in the northeast more efficient and increasing fuel economy standards?

In short, the article contains some useful analysis of where policies in the United States have been more beholden to the corporate interests that pushed them than to the national interest. However, looking at other countries, we see that Denmark is now producing more than 25% of its electricity from wind, and Germany is producing more than 15% of its electricity from renewables. Contrary to Everett’s statements, these efforts have paid off with major job increases in renewable energy production for wind, solar photovoltaic, and solar hot water.

The market and market decisions are not necessarily well matched to global politics, national security, environmental threats, or the long time frame associated with climate change. But if we are to rely on them at all, we should eliminate the many financial and indirect subsidies and externalities enjoyed by the fossil fuel industry.

WILLIAM R. MOOMAW

Professor of International Environmental Policy
Director, International Environment and Resource Policy
The Fletcher School
Tufts University
Medford, Massachusetts

Each year, the author and Bruce Everett debate energy and climate issues at the Fletcher School.

Bruce Everett details several important misconceptions and failures that undermine our energy innovation system. But his conclusion that the military should retreat from this endeavor ignores technologies that have the potential to both enhance military capability and thrive in the commercial market. In these instances, we can and should leverage the unique institutional attributes of our military, which as Everett points out catalyzed technological change in the 20th century. Biofuels and batteries demonstrate the need for a more nuanced perspective. Because alternative fuels do not make ships go faster or planes fly higher, the Navy has been criticized for their efforts to link biofuels R&D with national security priorities. Battery technologies, on the other hand, offer clear tactical and operational benefits to the Army and Marines.

The challenge then is to design programs that effectively move engineering knowledge between carefully aligned military and commercial applications. It is not to dissolve government/industry partnerships. The three phases Everett describes—conceptual, technical, and commercial—are eerily linear, and it is a contradiction to cite the successes of military innovation while exalting the role of basic research. The military is not Lewis and Clark–like, looking for nothing in particular. The military is looking for something it can buy, operate, and maintain. And it is precisely this end-to-end, incremental approach that fostered so much innovation in the 20th century. Faith in another approach, whether for defense or energy or health, is like faith in the Seven Cities of Gold. It is not enough to understand the essence of research. We must strive to cultivate the essence of innovation.

TRAVIS R. DOOM

Program Specialist
Consortium for Science, Policy and Outcomes
Arizona State University
Washington, DC

There is one serious and indeed fatal flaw in the article by Bruce Everett. It left out completely the great success story of energy-efficiency policies and technologies pursued over nearly 40 years, with the result of reducing the cost of many energy services to the consumer, often to the point of paying back the investment in a short period, much shorter than the lifetime of the technology purchased. In fact, the reduction in energy costs from only a few technologies supported by the Department of Energy (DOE) in concert with the private sector have repaid the entire government energy R&D budgets for energy efficiency from the beginning of DOE, as measured by dollars saved by the public. [See the 2001 National Research Council (NRC) report Energy Research at DOE: Was it Worth it?] Examples are low-e windows and high-efficiency ballasts for fluorescent lights. The same was true for fossil energy R&D. One particular innovation, the diamond incorporated drilling bit, has contributed significantly to the directional drilling so important to the current natural gas bonanza. A more recent study of the outlook for energy efficiency’s continuing role is given in the 2009 NRC report Real Prospects for Energy Efficiency in the United States. It estimated that cost-effective efficiency improvements could reduce U.S. energy use by 19 to 36 quads by 2030.

Another important point ignored by Everett is that overall energy use per unit of gross domestic product has steadily decreased. This reduction has been influenced by the offshoring of much of our manufacturing and the structural shift to a more service-based economy, but the biggest influence has been the use of more energy-efficient technologies and practices. From this, one can deduce that energy efficiency has contributed more to energy supply than have any of the supply-side technologies.

To evaluate the history of energy policies, one needs to look at the whole picture. Everett’s paper is fatally flawed, even if it had been titled “Back to Basics on Energy Supply Policy.”

WILLIAM FULKERSON

Senior Fellow
Institute for a Secure and Sustainable Environment
University of Tennessee
Knoxville, Tennessee

From 1975 through 2002, the U.S. government poured more than $10 billion in tax credits, research funds, and technology into a long-shot idea that commercial volumes of natural gas could be extracted economically from source rock. The effort was superlatively successful, triggering today’s shale gas and shale oil bonanza, which has transformed energy, balance of payments, and geopolitics for the United States. It has also especially benefitted one private player, a sole wildcatter with the patience and stubbornness to endure the years of uncertainty. That would be George Mitchell and his company, Mitchell Energy. In 2002, having absorbed and concentrated the lessons of the quarter-century of federally funded research into work in the Barnett Shale near Fort Worth, Texas, Mitchell sold his company to Devon Energy for $3.1 billion.

In his impressive article on how energy is best developed, Bruce Everett describes succinctly why private players almost always are best at divining the technologies that will actually work commercially. Advocates of robust government research efforts usually cite the Internet, GPS, and semiconductors as examples of what can be done. But, Everett writes, other high-priced government-funded triumphs—the Moon landing and supersonic flight among them—have yet to result in direct commercial application. The difference is that government funding often gets distorted by politics, whereas private players only have the bottom line in mind. Because of this history, Everett argues, the government should finance basic research and avoid joint efforts with industry, not to mention tax credits and renewable mandates. Above all, do not “pick winners” by supporting individual companies.

In his conclusions, Everett aligns with a broad swath of thinking. He is not alone in his advice. But this philosophy would hold up better if it defended itself in the context of the shale boom, which some call the greatest development in energy in a century.

If one had strictly observed Everett’s recommendations, the global economy would be in much worse shape, and the United States in particular would be, because oil and gas supplies would be much tighter. As for Mitchell, he would still be a billionaire from his previous ventures. But he would be at least a couple of billion dollars poorer. Some might say that, given the scale of the triumph when the government did step in effectively on behalf of a single player, picking winners is not necessarily a bad bet.

STEVE LEVINE

Schwartz Fellow, the New America Foundation
Washington correspondent, Quartz (http://www.qz.com)
Adjunct Professor, Energy Security
Georgetown University
Washington, DC

Bruce Everett has once again with clarity of vision and hard-hitting facts delivered a critique of U.S. energy R&D policy over the past 40 years that should be read by every high-ranking energy official in the administration, members of Congress, and the head of every energy trade association in Washington. With devastating detail, he shows how government financial support for the nuclear and renewable energy industries has cost Americans billions of dollars while doing little to change the country’s overwhelming dependence on fossil fuels. His assessments of the state of the U.S. solar, ethanol, and wind (especially offshore) industries are particularly stark, given the minimalist role that they play in the nation’s energy balance. Given this disappointing record, one has to ask why serious energy analysts continue to believe that we can have an energy future based predominately on renewables, demand-side management, and energy efficiency, when the facts clearly say otherwise. How long are we going to continue to hear the mantra that if the price of all fuels just reflected their “real” social and “environmental externalities,” we could make the conversion to a carbonfree future in the next 25 years? If this were so, why does every forecast for the next 25 years project global increases in fossil fuel consumption and rising CO2 emissions?

Where one has to quibble a bit with Everett, however, is in his singular lack of critique of the subsidies received by the fossil fuel industry, where he spent a large portion of his career. There is virtually no mention of the various tax loopholes or advantages for domestically produced oil or the favorable treatment of foreign-produced oil. Although this author has in the past written on how important some of these tax provisions are for independent oil and gas companies who find most of the oil produced in the United States each year, or how certain tax breaks on foreign-produced oil are necessary to level the playing field for U.S.-based international oil companies against their foreign competitors, these provisions are still subsidies little different from those received by the nuclear and renewable energy industries against which Everett rails. Likewise, he is singularly silent about the newest economic engineering tool (master limited partnerships) used increasingly to finance a host of energy projects, particularly natural gas and petroleum product pipelines, owing to their favorable tax treatment.

As a free-market economist, Everett obviously loathes interference in the market geared to forcing the commercialization of new technologies before they are able to compete in the marketplace on their own footing. Although this philosophical stance is respectable, why doesn’t he suggest that it is an equally viable proposition to get rid of all special tax advantages on all energy forms as well as all Clean Energy and Renewable Portfolio standards and put on a serious carbon tax and let the fight begin to see who can compete in the marketplace. At least we would get a reduction in CO2 and other greenhouse gas emissions, while keeping our offshore areas pristine and our food prices perhaps lower both at home and abroad.

CHARLES K. EBINGER

Senior Fellow and Director, Energy Security Initiative
Brookings Institution
Washington, DC

Changing science education

Carl Wieman’s “Applying New Research to Improve Science Education” (Issues, Fall 2012) brings a welcome and refreshing focus on science learning. Creating learning environments that support all students in developing expert-like thinking in science is essential, whether they will join the science and engineering workforce or rely on scientific approaches and findings in making health care choices or voting on land-use issues. As Wieman notes, we have the evidence in hand to improve K-16 learning, synthesized in the National Research Council reports Taking Science to School and Discipline-Based Education Research. Yet the evidence also reveals a striking lack of widespread implementation of teaching practices that take into account students’ prior knowledge and structure their learning with challenging but doable practice in solving problems like experts.

Gateway science, math, and engineering undergraduate courses are key to improving K-16 learning. These courses model science learning and teaching for future teachers and create impediments for more than half the students who enter college intent on a science or engineering major, but leave science because they are discouraged by the poor teaching. The new framework and standards for science education are firmly rooted in the evidence on science learning, but they will be difficult to implement fully if future teachers do not experience these evidence-based practices in their own undergraduate years. Further, all imaginable success in improving K-16 science learning will be for naught if high-school graduates enter traditional lecture-based college courses. The new standards can motivate change in the gateway courses, but this alone is not sufficient.

For the many reasons outlined in Wieman’s article, university culture works against improving teaching. Changing the incentive and rewards systems to create departmental and college-wide cultures that value and recognize effective teaching has been a Sisyphean task. Certainly policy levers focused on accountability are one way to push on the interconnected system of teaching practice, curriculum development, and assessment. Bottom-up efforts from faculty alone cannot leverage the scale of change required. Yet it is faculty behavior that must change. The typical professional development approaches of sharing “what works” and offering evidence has been quite successful in raising awareness of instructional strategies that support science learning, but a relatively small percentage of faculty persist in using effective strategies. Lack of understanding of the principles behind a practice or of how to adapt the practice to one’s specific context are common barriers to persisting with a practice. We need to find ways to both help and provide incentives to faculty to put effective practices into place.

Lasting change needs to be at the level of departments and institutions, achievable only by applying multiple levers, including policy aimed at the incentive system and new forms of professional development. National efforts, including programs by the National Science Foundation, National Institutes of Health, and Howard Hughes Medical Institute Partnership in Undergraduate Life Science Education, model creative ways stakeholders can partner to push change at scale. However, as Wieman indicates, implementation must be anchored in the research findings on science learning.

SUSAN RUNDELL SINGER

Laurence McKinley Gould Professor of Biology
Department of Biology
Carleton College
Northfield, Minnesota

The author chaired the National Research Council’s Discipline-Based Education Research committee.

Measuring science

The excellent article “Qualitative Metrics in Science Policy: What Can’t Be Counted, Counts” by Rahul Reki and Neal Lane (Issues, Fall 2012) reminds me of a situation that Neal Lane and I lived through during the time he was director of the National Science Foundation (NSF) and I was chair of the National Science Board. We had heard that an Ohio congressman was likely to vote against the NSF budget. Neal had the opportunity to meet with him in his office. I do not remember for sure whether I was along. The congressman asked whether it might be a fair estimate to believe that only one-third of the funds NSF gave led to research having societal impact. One-third might be an overgenerous estimate. Maybe it was more like one-tenth. The congressman said that, granting him this hypothesis, he believed the NSF budget should be cut by two-thirds! Some moments of unease followed. It was agreed that to spend taxpayer dollars having no societal consequences was a waste. The problem was determining which projects had societal benefits.

For example, in the late 1930s, a few people such as Rabi at Columbia University were trying to measure the energy-level structure of atoms having nuclear spin in a magnetic field. These energy spacings were minuscule, much smaller than the energy of thermal fluctuations at room temperature. No one could imagine at the time what practical consequences these radiofrequency experiments on atomic beams could possibly have. Later, Bloch at Stanford and Purcell at Harvard would independently discover that different atomic nuclei within a molecule resonate at different radiofrequencies for the same magnetic field strength, allowing us to learn chemical and structural information about the molecule. Still later, in the 1970s people such as Lauterbur at the University of Illinois and Mansfield at the University of Nottingham would show that a magnetic field whose strength changes over space allowed images to be made. This ushered in what is called magnetic resonance imaging, which is used to distinguish pathologic tissue (such as bone breaks or tumors) from normal tissue. It is why today doctors know how to fix various bone fractures without first cutting you up to see what is wrong.

The congressman understood and did vote for the NSF budget. The real point of this story, however, is to emphasize what Reki and Lane have already stressed: A smart science policy is not simply based on enumeration.

RICHARD N. ZARE

Department of Chemistry
Stanford University
Stanford, California

In reading the article by Rahul Reki and Neal Lane, I was reminded of Lord Kelvin’s famous remarks of a century ago that have had an enormous impact on thinking about what a strong field of study ought to be like. “When you can measure what you are speaking about, and express it in numbers, you know something about it; but when you cannot measure it, when you cannot express it in numbers, your knowledge is of a meager and unsatisfactory kind.” However, this proposition implicitly denies the fact that much of solid human knowledge about the world is not expressed in numbers. Much of scientific knowledge is not quantitative. You and I agree that the tree we see whose leaves are turning red is a maple tree. Only a very small portion of Darwin’s description of species in his great book, or of the description of phenomena addressed by geology, is expressed quantitatively. These scientific fields could be even stronger if some of the phenomena that presently are described qualitatively had a strong quantitative characterization. But we still know a lot about many phenomena that don’t. Kelvin’s argument is greatly exaggerated.

And thinking of this kind has led in a number of fields to a messianic effort to construct measures or indicators of the things being studied, together with a position that the science ought only to be about the phenomena that are characterized quantitatively. This raises two kinds of questions. One is how good particular quantitative indicators are in characterizing what we are really interested in understanding. The other is whether the range of good quantitative indicators we have covers what we want to know.

These issues are presently playing out dramatically regarding the evaluation of what students are learning in school and the quality of the teachers who are trying to help those students learn. Nobody is arguing that it would not be convenient if we had good quantitative measures of both. The argument is about whether the indicators we have and use can adequately characterize what we want to know about.

This is exactly the issue regarding the insistence that a rate–of-return calculation be made regarding aspects of academic science. Many years ago, in my first testimony before a congressional committee, I was asked what the rate of return was on university research. I answered that we did not know, that calculations of that could be all over the map, and that all the computational algorithms that I knew about were focused on only a small portion of how scientific research influences the human condition. That still strikes me as right. A good portion of what is important is not readily quantitative, indicators that are developed are likely to miss much of what we are interested in, and at the same time a considerable amount can be learned about those things by qualitative observation and analysis.

My argument, of course, is not that we shouldn’t try to develop quantitative indicators. But we should resist becoming their servants.

RICHARD NELSON

Columbia University
New York, New York

The article by Rahul Rekhi and Neal Lane (no relation) was very thoughtful and made a number of important points about science measurement. Indeed, they are almost exactly the same points that the Office of Science and Technology Policy Science of Science Policy interagency group made in our 2008 Roadmap, in which we concluded that “the data infrastructure is inadequate for decisionmaking.”

Rekhi and Lane are exactly right: There is increased pressure to document the results of federal investments regardless of whether those investments are in education, welfare, foreign aid, health, workforce, or science. The scientific community should take this pressure seriously; the American Recovery and Reinvestment Act (ARRA) reporting requirements are a likely harbinger of future requests. And scientists who landed men on the Moon and mapped the human genome should be the last group to argue that measurement is too hard to do or should not be done, particularly when there is ample literature that spells out how to do such measurement in a scientific manner in many other contexts. I will illustrate with comments on four of the key points made by Rekhi and Lane.

(1) Much of the benefit associated with science is in training students. Rekhi and Lane make the argument that it is “intuitive” that research opportunities provided to undergraduates have value, and cite the National Science Foundation’s (NSF’s) Research Experiences for Undergraduates program. The 2012 budget for this program was $68 million. Intuition might well be supported by evidence. That evidence does not need to be in higher test scores or salaries. The point is made that it is “a critical aspect of the educational process of becoming a truly 21st-century–ready scientist or engineer.” How might that be determined? Here are some obvious and I hope noncontroversial steps fully in the spirit of Louis Pasteur’s swan flask experiment.

Define the population. NSF and other science agencies should be able to tell the public exactly how many students are supported by science funding at all levels. Our experience with ARRA reporting was that universities and science agencies were unable to provide that information in a consistent, reliable manner. Write down the outcome measures. What does it mean to be a truly 21st-century–ready scientist and engineer? These can be qualitative or quantitative, but we should be able to enunciate them clearly. Write down a theory of change. What do scientists think is the causal effect of the program on the outcome of interest? Write down a counterfactual. What would the results be if the program did not exist? This then leads to the next point.

(2) It is difficult to quantitatively measure the results of science. True. It is difficult to measure almost everything (including the human genome). But scientists use both quantitative and qualitative data routinely to make judgments about other scientists: in tenure and promotion decisions, whom to include in conferences, who is “good” and not good. Write down what those data are.

(3) The benefits are not just economic. This argument has not been made by the science of science policy community. In fact, the concern of that community is that if scientists do not develop their own scientifically grounded measures, the only measures that will be used will be economic, as we saw with ARRA reporting.

(4) The benefits take a long time to accrue. That is clearly the case, but is no less true than of investments in education, health, transportation, and many other areas of federal investment. The processes by which scientific ideas are created, transmitted, and adopted are as well understood and well studied as in any other field of human endeavor. Numerous trace studies have described the network processes in great detail, and the fact that the processes are not deterministic does not mean that there are no covariates (see points 1, 2, and 3 above).

Einstein was right; not everything can be counted. But we know better as scientists to take that argument to its full reductio ad absurdum extent. Scientists can, and do, attempt to measure almost all aspects of the universe. Science itself should be no exception.

JULIA LANE

Senior Managing Economist
American Institutes for Research
Washington, DC

Managing water

Water is perhaps the world’s most vital natural resource. All humans need clean fresh water for drinking, cooking, and washing. Modern society depends on adequate water supplies for agriculture and industry, fisheries and forestry, to generate power, and to eliminate wastes. As a result, the possible forms and motivations of potential social tensions and political frictions over water management are as varied as the societal benefits that are supplied by water.

Examining the links between water and social conflict, Ken Conca (“Decoupling Water and Violent Conflict,” Issues, Fall 2012) rightly emphasizes that the greatest risks of large-scale violence stem not from actual scarcity of water but from ineffective, illegitimate, ill-adapted, or even absent institutional arrangements to govern water, and that organizational structures and policy mechanisms that narrowly define the problem as physical water supplies ill-prepare decisionmakers to sustainably manage water resources.

Establishing and empowering effective and accepted water institutions, however, can be a herculean task. Freshwater sources and flows, river basins, groundwater aquifers, and lakes ignore political and bureaucratic boundaries. Yet responsibilities for managing water are typically divided internationally between riparian countries and fragmented domestically between rival agencies, often representing different interests. For the past two decades, much of the global water policy community has striven to tackle these challenges by developing strategies for integrated water resources management (IWRM). IWRM seeks to balance supply and demand dynamics, coordinating between multiple uses, constituencies, and ecosystem needs, as well as across geographic areas. IWRM recognizes water as both a social and an economic good, and so promotes participatory policy approaches engaging stakeholders at all levels in an effort to manage water resources both efficiently and equitably.

Despite this conceptual promise, however, many observers fear IWRM has proven more successful as an incantation than in implementation. Yemen, one of the world’s most water-stressed countries, provides a cautionary case in point. As recently as 2006, the fourth World Water Forum lauded Yemen for incorporating IWRM tenets into the letter of its national water policy. Behind this notional commitment, though, relevant agencies lack the human and technical capacities to administer and enforce water policy. Powerful vested interests in Yemeni state and society often oppose IWRM approaches. Perverse subsidies, for example, on diesel fuel for well pumps, have abetted the rampant expansion of groundwater irrigation. Critically, groundwater furnishes 70% of total water withdrawals in Yemen, but the country is now depleting its aquifers two to four times faster than nature can replenish them. At that rate, a 2010 World Bank assessment concluded that Yemen will essentially exhaust its groundwater reserves by 2025–2030. Although Yemen embraces IWRM in principle, in practice the country stands on the brink of water crisis.

Around the world, recent regional and global status reports reveal that many envisaged IWRM reforms have yet to be fully implemented, while progress in essential areas such as environmental monitoring and integration, climate change adaptation, stakeholder participation, knowledge-sharing programs, and sustainable financing is lagging. Meeting the world’s growing water needs will require rectifying these shortcomings. To ensure sustainable supplies of water, policymakers must ensure the supply of capable institutions.

DAVID MICHEL

Director, Environmental Security Program
The Stimson Center
Washington, DC

Ken Conca’s essay provides a sweeping overview of many issues pertinent to the global water situation. Unfortunately, it contains a number of internal inconsistencies and contradictions that leave the reader wondering what solutions Conca may actually be offering, or whether he believes that water and conflict are, in fact, coupled.

For instance, Conca cites a recent African study that found that ”deviations from normal rainfall patterns led to an increase in social conflict, including demonstrations, riots, strikes, antigovernment violence, and communal conflict between different social groups.“ He cites results from another study suggesting that water-related conflicts occur in the Middle East/ North Africa region every 2.5 days on average. Those studies seem to provide considerable evidence that water shortages can result in conflict. Yet Conca concludes that “efforts to correlate water scarcities or drought with the onset of civil war have for the most part not found a statistically significant link” and “If there is a risk of large-scale violence, it stems not from physical water scarcity but from the institutionalized arrangements that shape how actors govern water.” If Conca is suggesting that inadequate governance and not physical scarcity is at the root of water conflict, then he has inappropriately placed the governance cart before the water horse.

Some clarification and reinterpretation are warranted here. It is painfully clear that physical water scarcity—a result of water consumption approaching or exceeding the limits of water availability—can be a potent ignition source for conflict among those sharing the water resource. When a watershed community is collectively using more water than the watershed can bear, trouble and conflict are close at hand. It is also clear, given great disparities in water availability and water demands across the borderlines of local watersheds, that water conflicts are intensely localized in their origins. At the exact same moment, one watershed can be experiencing severe scarcity while an adjacent watershed has sufficient water to satisfy all demands. The primary issue with governance, then, is a rather personal one: Are we managing our demands within the limits imposed by local water availability?

With my son beginning his first semester of college, I am finding it exceedingly difficult not to draw parallels between the basic skills of managing a personal checking account and managing a local water resource. I have advised my son that he will experience all sorts of temptations to spend more than his monthly budget will allow, and that there will be plenty of surprises like parking tickets and library fines to pay. But ultimately, if my son or my local watershed community ends up in overdraft, they should not lay blame on the bank or their government or other external parties for causing the problem. We must look first to our own internal conflicts over our individual or collective lack of discipline for managing ourselves in ways that avoid scarcity and enable us to be highly functioning citizens of the world.

BRIAN RICHTER

Director, Global Freshwater Strategies
The Nature Conservancy
Crozet, Virginia

We’re not dopes

I share much of the communitarian emphasis and values that Amitai Etzioni has capably expressed in numerous writings. I also share Etzioni’s critical stance toward rational choice theory. That said, I find his article “The Limits of Knowledge: Personal and Public” (Issues, Fall 2012) disappointing.

The article begins by noting that one of the basic assumptions underlying much of Western thinking is that individuals are rational beings. Etzioni then introduces the relatively new field of behavioral economics, arguing that this field “has demonstrated beyond reasonable doubt that people are unable to act rationally and are hardwired to make erroneous judgments.”

I think Etzioni is too eager to throw the baby out with the bathwater (although imperialist rational choice theorists might need to take a bath of bounded rationality). The many laboratory experiments he cites of Kahneman, Thaler, and Ariely of course indicate limitations on the pure tenets of rational choice theory. When it comes to “real-life” studies, it is harder to interpret a finding: So on page 53, the case of Israeli parents arriving late to pick up their children at a daycare center with a 10-shekel fine might well be a rational action of having decided that their time was worth more than the price of the fine; similarly for the next example of the majority of participants turning down an annuity (depending on the costs of the annuity, a lump sum payment might be more advantageous). Too much of Etzioni’s arguments seem to view actors as cognitively limited cultural dopes with “limited capacity to digest data” (p. 54). I doubt that any behavioral economist would argue that humans only act irrationally.

I don’t see how with this rather bleak perspective, his benevolent communitarianism can hope for an intellectual shift “of a Copernican magnitude” (p. 55). I would highly recommend that Etzioni consider as a needed supplement the vigorous new economic thinking of 1998 Nobel Prize winner Amartya Sen and especially 2009 Nobel Prize winner Elinor Ostrom for her emphasis on developing “a more general theory of individual choice that recognized the central role of trust in coping with social dilemmas … The frameworks and empirical work that many scholars have undertaken in recent decades provide a better foundation for policy analysis.” That, it seems to me, is a more promising macro-oriented thrust to give new vigor to communitarian thinking.

EDWARD A. TIRYAKIAN

Professor Emeritus of Sociology
Duke University
Durham, North Carolina

Energy and human behavior

In “What Makes U.S. Energy Consumers Tick?”, Kelly Gallagher and John Randell (Issues, Summer 2012) succinctly summarize a wide range of fundamental and applied research questions to which policymakers, consumers, governments, and energy producers/distributors will need clear answers, if the nation is to break from unsustainable and environmentally detrimental energy consumption habits.

The authors articulate key research questions and through their overview (1) highlight the immense and complex role of human behavior in energy production and consumption; (2) identify the critical need to better understand the most fundamental influences on individual, group, and societal behavior; (3) underscore science findings showing that behavior is inextricably integrated with economic prosperity, technology, and the health of civil society; and (4) emphasize the behavioral and economic implications for the shelf life of our standard of living and prospects for improving that of future generations.

Gallagher and Randell identify a monumental agenda for industry, policymakers, and science, generally. I say science, generally, because grand challenges such as these require “convergent science,” integrated responses across the full range of sciences; the challenges cannot be solved within one disciplinary framework. Diverse applied and interdisciplinary domains of physical, behavioral, and social research are necessary to ensure that we understand how we can take collective control of our energy practices.

Gallagher and Randell, however, do not address one critical issue affecting our ability to tackle this research agenda. Specifically, the community of behavioral and social scientists who are trained and interested in tackling these research questions is quite small. Our talent reservoir is shallow. A colleague at the November 2012 workshop on “Integrating Social and Behavioral Energy Research Activities,” organized by the American Academy of Arts and Sciences, lamented facetiously that the number of research topics alone almost outstrips the number of current researchers. But the National Science Foundation (NSF) is helping expand this community through a number of innovative interdisciplinary programs. For example, the cross-agency SEES (Science, Engineering, and Education for Sustainability) initiative draws on every part of NSF’s research and education portfolio and engages physical, social, and behavioral sciences. NSF programs such as SEES Fellows, Dynamics of Coupled Natural and Human Systems, Sustainability Research Networks, and Interdisciplinary Research in Hazards and Disasters are developing the critically important interdisciplinary scientific talent pool.

NSF is also engaging the President’s Council of Advisors on Science and Technology, the Department of Energy, and the National Oceanic and Atmospheric Administration in discussions on reducing technical and behavioral barriers to energy efficiency, while simultaneously helping to build convergent research communities and agendas through its support of SEES Research Coordination Networks and workshops. Numerous NSF programs support interdisciplinary energy-related research, training, and team development through graduate student support, Research Experiences for Undergraduates, INSPIRE (Integrated NSF Support Promoting Interdisciplinary Research and Education), the Science of Organizations, Decision Making Under Uncertainty, and many other programs that play a critical role in addressing the human resource shortfall.

In this era of “big data” (for example, from electricity “smart meters”), a growing community of interdisciplinary social and behavioral scientists can develop a sound understanding of effective levers by which we can control our energy use and costs (personal and environmental), providing robust market-governed options that enable Americans to create their own energy future.

The energy-related behavioral and social science research agenda outlined by Kelly and Randell could well pave the way for a 21st century convergent science knowledge base that can inform effective policies to address other grand challenges (for example, waste generation and management and crime prevention). In addition, this agenda could serve as a model for the many domains in which an understanding of behavioral drivers, modulators, and influences is critical to the functional health, sustainability, and advancement of modern living standards.

CORA MARRETT

Deputy Director
National Science Foundation
Arlington, Virginia

History Lesson

When you’re watching Steven Spielberg’s terrific new movie Lincoln, remember that this was the same Congress that passed the Morrill Act and created the National Academy of Sciences (NAS). Although many of them clearly harbored deeply racist beliefs and were not above narrowly self-interested politicking, they were also capable of tackling profound questions of human rights and the growing importance of science and education. Even in the midst of a battle for the nation’s survival, they were able to look beyond their current situation and maintain an ambitious vision of the future.

The contrast to today’s Congress is painful to see. It required an enormous effort to postpone the crisis that was created by its own decision a few months earlier to construct a fiscal cliff. And when members of Congress should be turning their attention to a long-term fiscal plan that balances government income, entitlements, and discretionary spending, we are likely instead to be subjected to a pointless debate about the debt ceiling in which our elected leaders ponder the difficult question of whether the nation should pay the bills it has already incurred.

Still, we should temper our praise for 19th century wisdom. As historian Daniel Kevles reveals in his enlightening article in this issue, the creation of the NAS did not mark the beginning of an enlightened age. Many members of Congress gleefully mocked the notion that scientific eggheads could provide practical guidance on important political questions, and many people preferred to be guided by what they wished to be true rather than what evidence and analysis indicated was reality. This archetypal drama has played out many times in the past 150 years, and it will certainly provide rueful entertainment in the future.

What is remarkable, nevertheless, is that the United States created an institution designed to incorporate scientific knowledge into government decisionmaking. And in spite of waves of reaction against specific scientific findings and an ever-present strain of anti-intellectualism, or at least anti-elitism, in the American character, the influence of the scientific community and of the NAS has grown. It certainly helps that science and technology have contributed to stunning advances in public health and economic productivity, but it also matters that some scientists, engineers, and physicians have accepted the responsibility to play a role in public debates and to accept roles of national leadership.

The NAS will be holding many commemorative events this year, and Daniel Kevles and others will be completing a history of the institution. This history includes much to be proud of, but we should also take this opportunity to learn from experience. The institution itself has grown, expanding from science to include engineering, medicine, and the social sciences. As science and technology have become increasingly prominent components of modern life, the scientific community has begun to understand how its members must also become more fully integrated into all aspects of civic life. No doubt there are other lessons we have learned and need to learn. Let’s make this year of celebration also a year of contemplation.

Pathways to the Middle Class: Balancing Personal and Public Responsibilities

The United States defines itself as a nation where everyone has an opportunity to achieve a better life. Correspondingly, everyone should have the opportunity to succeed through talent, creativity, intelligence, and hard work, regardless of the circumstances of their birth. This meritocratic philosophy is one reason why U.S. residents have customarily shown relatively little objection to high levels of economic inequality—as long as those at the bottom have a fair chance to work their way up the ladder. Similarly, people are more comfortable with the idea of increasing opportunities for success than with reducing inequality.

It is critical, then, to identify and employ effective and affordable ways to provide children from every race and every social and income group with ample opportunities to succeed. Here, a considerable body of research points to a variety of promising interventions at various stages of life, from before birth through entry into adulthood.

Who succeeds and why?

One of the ways of thinking about opportunity is in terms of generational improvement in living standards. Among today’s middle-aged population, four in five households have higher incomes than their parents had at the same age, and three in five men have higher earnings than their fathers. The extent to which this will be true for today’s children remains to be seen. More importantly, if everyone grows richer over time, but the economic fates of individuals are bound up in their family origins, then in an important sense opportunities are still limited. If poor children have little reason to believe that they can grow up to be whatever they want, it may be of little comfort to them that they will probably make more than their similarly constrained parents. A better-off security guard may still have wanted to be a lawyer.

The reality is that economic success in the United States is not purely meritocratic. People do not have as much equality of opportunity as society would like to believe, and the United States offers less mobility than some other developed countries. Although cross-national comparisons are not always reliable, the available data suggest that the United States compares unfavorably with Canada, the Nordic countries, and some other advanced countries.

People do move up and down the ladder, over their careers and between generations, but it helps if you have the right parents. Children born into middle-income families have a roughly equal chance of moving up or down once they become adults, but those born into rich or poor families have a high probability of remaining rich or poor as adults. The chance that children born into families in the top income quintile will end up in one of the top three quintiles by the time they are in their forties is 82%, whereas the chance that children born into families in the bottom quintile will do so is only 30%. In short, a rich child is more than twice as likely as a poor child to end up in the middle class or above, according to data presented in 2012 by the Pew Economic Mobility Project.

Why do some children do so much better than others? And what will it take to create more opportunity?

In order to better understand the life course of children, especially those who are disadvantaged, we have developed, along with colleagues at the Brookings Center on Children and Families, a life-cycle model called the Social Genome Model. The model divides the life cycle into six stages and specifies a set of outcomes for each life stage that, according to the literature, are predictive of later outcomes and eventual economic success. The life stages are family formation, early childhood, middle childhood, adolescence, transition to adulthood, and adulthood. The success indicators include being born to parents who are ready to raise a child, being school-ready by age 5, acquiring core competencies in academic and social skills by age 11, being college- or career-ready by age 19, living independently and either receiving a college degree or having an equivalent income by age 29, and finally being middle class by middle age.

The data we use, from the National Longitudinal Survey of Youth, follow children born primarily in the 1980s and 1990s, starting in 1986 through 2010. We have projected their adult incomes using a statistical model. The projections in childhood and adulthood in our final data set closely track estimates from independent sources such as the U.S. Census Bureau and other surveys.

Among our findings, 60% of the children we followed into adulthood (through age 40) live in a family with income greater than 300% of the poverty line set by the Census Bureau (about $68,000 for a married couple with two children). In short, they have achieved the middle-class dream. This means, however, that roughly 40% have not achieved this success. The reasons are many, but by looking at where children get off track earlier in life, it is possible to begin to see the roots of the problem.

As our data show, about two-thirds or more of all children get through early and middle childhood with the kinds of academic and social skills needed for later success. However, a large portion of adolescents falls short of achieving success even by a relatively low standard. Only a little over half manage to graduate from high school with a 2.5 grade point average during their final year (the only period for which we have data on grades) and have also not been convicted of a crime or become a parent before age 19.

Success is nevertheless more common a decade later. Sixty percent of children will live on their own at the end of their 20s, either with a college degree in hand or with family income of approximately $45,000, which is about 250% of the level at which a family is considered poor. It is roughly equivalent to what a college-educated individual of the same age can expect to earn working full time, according to a 2011 report by the U.S. Department of Education. People in their late 20s without a college degree might achieve this college-equivalent income through on-the-job training, living in a dual-earner household, or by other means.

Putting the adolescent and adult results together, it appears that F. Scott Fitzgerald was wrong. There are second acts in American lives; lots of people make it to the middle class in adulthood despite entering life inauspiciously. But they would almost certainly be better off if they had acquired more skills at an earlier age, and especially if they had navigated adolescence more wisely.

Another way to give meaning to the success probabilities is to look at how success varies by different segments of the population. Consider gender. Boys and girls enter the world on an equal footing, but they take different paths on the way to adulthood. By age 5, girls are much more likely than boys to be academically and behaviorally ready for school. That advantage persists into middle childhood and adolescence. Men catch up in early adulthood and then surpass women in terms of economic success. At age 40, 64% of men but just 57% of women have achieved middle-class status.

The finding that girls do better than boys during the school years is not new. Girls mature earlier than boys and are better able to sit still and follow directions, and thus benefit more from classroom learning at a young age. A number of studies have shown that girls are less likely to act out, drop out of school, or engage in behaviors such as delinquency, smoking, and substance abuse, even as they are more likely to experience depression and face the risk of a pregnancy during adolescence.

Women are not only much more likely to graduate from high school, but they now earn 57% of all college degrees and more graduate degrees, as well. Despite their educational advantages, once they are adults, women still earn less than men. Although this earnings gap has declined sharply during the past half century, women who work fulltime still earn about 80% of what men earn, according to the U.S. Department of Commerce. The reasons for their lower earnings are related to the fact that, mainly as a result of their family responsibilities, women’s labor force participation and hours worked are still lower than men’s, and they are concentrated in occupations that pay less than those held by men. The extent to which these occupational differences reflect social constraints on women’s roles versus their own preferences has been hard to sort out. Single parenthood also accounts for some of the gender gap in adulthood. Single mothers are much more common than single fathers. Even if they work, many single mothers disproportionately bear the burden of feeding additional mouths with their paycheck as compared with noncustodial fathers, whose child support is often modest.

Other patterns of inequality are much stronger. In our study, 68% of white children are school-ready at kindergarten, compared with 56% of African American children and 61% of Hispanic children (although data on Hispanic children were limited). In middle childhood, black children fall farther behind, and by adolescence, both African Americans and Hispanics are far behind white children. By age 30, both are still lagging, although Hispanics have done some catching up. By age 40, a whopping 33–percentage-point difference between blacks and whites persists. Only one-third of African Americans have achieved middle-class status by middle age. Even among those who do reach the middle class, their children are much more likely to fall down the ladder than the children of white middle-class families. The black-white gaps, and to a lesser extent the Hispanic-white gaps, are sizable for both boys and girls, although they are bigger among boys at the start of school and at the end of the high-school years.

If success at each stage varies by gender and race, it varies even more by the income of one’s parents. Only 48% of children born to parents in the bottom quintile of family income are school-ready, compared with 78% of children in the top quintile at birth. The disparity is similar in middle childhood. Perhaps most stunning, only one in three children from the bottom quintile graduates from high school with a 2.5 grade point average and having not been convicted or become a parent. For children from the top quintile, this figure is 76%. Parental income also affects the likelihood of economic success in adulthood, with 75% of those born into the top quintile achieving middle-class status by age 40 versus 40% of those born into the bottom quintile.

Finally, we have examined success rates for children who are born into more or less advantaged circumstances, defined more broadly than by their income quintile. Among children born of normal birth weight to married mothers who have at least a high-school education and who were not poor at the time of the child’s birth, 72% can be expected to enter kindergarten ready for school, whereas the rate is only 59% for women outside of this group. This gap never narrows, and by the end of adolescence, children with less advantaged birth circumstances are 29 percentage points less likely to succeed. At age 40, there is a 22–percentage-point gap between these groups of children in the likelihood of being middle class.

Children born advantaged also retain a large advantage at the end of the next life stage, early childhood. The same pattern prevails for subsequent stages: success begets later success. In middle childhood, adolescence, and adulthood, those who succeeded in the previous stage are much more likely than those who did not to succeed again. For example, 82% of children in our sample who entered school ready to learn mastered basic skills by age 11, compared with 45% of children who were not school-ready. Acquiring basic academic and social skills by age 11 increases by a similar magnitude a child’s chances of completing high school with good grades and risk-free behavior, which, in turn, increases the chances that a young person will acquire a college degree or the equivalent in income. Finally, success by age 29 doubles the chances of being middle class by middle age. In short, success is very much a cumulative process. Although many children who get off track at an early age get back on track at a later age, and can be helped to do so, these findings point to the importance of early interventions that keep children on the right track.

These features of social mobility processes are captured in Table 1, which shows the chance of becoming middle class for eight paths through the transition to adulthood. These are the paths taken by 76% of the children in our sample. The table shows that success in all four stages before adulthood is actually the most common pathway for children to take between birth and age 29, with over one-quarter of children taking that route and 81% of them achieving middle-class status. Of those who fail in all four life stages, a group that is only 8% of our sample, just 24% become middle class by age 40.

Table 1 also shows how early success or failure can matter even among those succeeding in the transition to adulthood. If children consistently fail before succeeding at age 29, they have a 60% chance of being in the middle class at age 40. If they consistently succeed in the earlier stages, they have an 81% chance. Similarly, people failing during the transition to adulthood have only a 24% chance of making it to the middle class if they have a consistent history of failure, but a 51% chance if they have a consistent history of success.

TABLE 1

Probability of reaching the middle class by number of successful outcomes and possible pathways

TABLE 2

Probability of reaching the middle class by number of successful outcomes and possible pathways through adolescence

On the other hand, early failures need not be determinative if children can get back on track. Children who are not school-ready have a similar chance of being middle class as those who are school-ready if they can get on track by age 10 and stay on track. Indeed, a striking feature of Table 1 is how well success at age 29 predicts success at age 40. This may be an artifact of the way we define success at age 29. Because it includes having an income of 250% of the poverty level, individuals need only increase their income by about 20% over the next decade to become middle class by middle age.

For this reason, in Table 2 we ignore the transition to adulthood and consider how well particular paths through age 19 predict middle-class status at age 40. These six paths through adolescence cover 88% of the children in our sample, and the two paths involving three consecutive successes or failures account for nearly half of all children. At first glance, success or failure in early childhood seems less important than in subsequent stages. Children who succeed in all three stages have a 74% chance of being middle class at age 40, whereas those who are not school-ready but succeed in middle childhood and adolescence have a 71% chance—essentially no difference. Comparing paths 2 and 4, which are identical except for early childhood, gives a similar impression.

It would be wrong to conclude, however, that early childhood success is unimportant. If the primary way that school readiness affects the likelihood of becoming middle class is by directing people into more successful paths, that fact would be obscured by comparisons such as these. If a child is school-ready, there is a good chance that he or she will continue to succeed in later stages. If a child is not ready, it is relatively unlikely that he or she will get on track. Paths 1 and 2 are much more common than paths 3 and 4; recovery from early failure is relatively rare.

To summarize the data further, we calculated the probability of reaching the middle class by middle age based on the number of life stages in which the individual experienced success. As already noted, an individual who experiences a successful outcome at every life stage has an 81% chance of achieving middle-class status. The probability of achieving this status decreases with each additional unsuccessful outcome. An individual who hits just one speed bump has a 67% chance of reaching the middle class, but this figure drops to 54% if there are two unsuccessful outcomes, 41% with three unsuccessful outcomes, and 24% for those who are not successful under our metrics at any earlier stages in life.

Who are the children who succeed throughout life? They are disproportionately from higher-income and white families, whereas those who are never on track are disproportionately from lower-income and African American families. For example, 19% of children from top-quintile families stay on track throughout their early life, while only 2% of children from bottom-quintile families do. The figures by race show that 32% of white children but only 10% of black children stay on track.

The fact that nearly one-quarter (24%) of our off-track individuals (those failing at every one of our metrics) manages to achieve middle-class status is a reminder that there are quite a few individuals who are economically successful despite a lack of success during school. Such individuals may be successful because of skills or personality traits that are not captured by our metrics, or by relying on support from another family member, or purely by luck.

Unfortunately, birth circumstances are highly predictive of the likelihood of achieving success in the four life stages preceding middle age. Children born at a low birth weight or who have a mother who is poor, unmarried, or a high-school dropout—circumstances we denote as less advantaged—have only a 17% chance of achieving all four interim markers of success. Figure 1 shows the likelihood that an individual achieves success at any particular number of life stages conditional on their circumstances at birth.

Although few children from less advantaged backgrounds succeed at every life stage through their 20s, it appears that when they do, they are nearly as likely to reach the middle class as children born into more advantaged families. Figure 2 indicates that children from less advantaged families who stay on track have a 75% chance of joining the middle class, compared with 83% for their more advantaged peers. On the other hand, although additional successes do increase the chances of ending up middle class for disadvantaged children, they often increase the chances more for advantaged children. Advantaged children are about as likely to end up middle class as disadvantaged children who have an additional success under their belt. For example, an advantaged child with no successes through their 20s is nearly as likely to end up middle class as a disadvantaged child with one success. Family background matters because it tends to affect childhood success at each stage of the life cycle, but also because it affects the likelihood of ending up middle class even for children experiencing similar trajectories through early adulthood.

FIGURE 1

Number of successful life stages by circumstances at birth

FIGURE 2

Chances of reaching the middle class by number of successful life stages

Help on the ladder

Because a number of studies have demonstrated that the gaps between more and less advantaged children are large and getting larger, without a clear plan to prevent the less advantaged from getting stuck at the bottom, the nation risks developing a society permanently divided along class lines. But our results and those of other researchers reveal some good news: Although it may be difficult for many children to stay on track, especially disadvantaged children, birth circumstances do not have to become destiny.

When children stay on track at each life stage, adult outcomes change dramatically. Parents who tell their children to study hard and stay out of trouble are doing exactly the right thing. But government can play a role as well, even if limited. In particular, government at various levels can adopt policies to foster intervention programs targeting every life stage to help keep less advantaged children on track. The research community has now identified, using rigorous randomized controlled studies, any number of successful programs, some of which pass a cost/benefit test even under conservative assumptions about their eventual effects. The following descriptions of successful programs are illustrative, not exhaustive.

Family formation. The first responsibility of parents is to not have a child before they are ready. Yet 70% of pregnancies to women in their 20s are unplanned, and, partly as a consequence, more than half of births to women under age 30 occur outside of wedlock. In the past, most adults married before having children. Today, childbearing outside of marriage is becoming the norm for women without a college degree. To many people, this is an issue of values; to others, it is simple common sense to note that two parents are more likely to have the time and financial resources to raise a child well. Many young people in their 20s have children with a cohabiting partner, but these relationships have proven to be quite unstable, leading to a lot of turmoil for both the children and the adults in such households.

Government can help to ensure that more children are born into supportive circumstances by funding social-marketing campaigns and nongovernmental institutions that encourage young people to think and act responsibly. It can also help by providing access to effective forms of contraception, and by funding teen pregnancy prevention efforts that have had some success in reducing the nation’s high rates of very early pregnancies, abortions, and unwed births. A number of well-evaluated programs have accomplished these goals, and they easily pass a cost/benefit test and end up saving taxpayers money.

Early childhood. The early childhood period is critical, and home visiting programs and high-quality preschool programs, in particular, can affect readiness for school. Home visiting programs, such as the highly regarded Nurse Family Partnership, focus on the home environment. Eligible pregnant women are assigned a registered nurse who visits the mother weekly during pregnancy and then about once a month until the child turns age 2. Nurses teach proper pre-and postnatal health behaviors, skills for parenting young children, and strategies to help the mother improve her own prospects (family planning, education, work). The program is available to low-income first-time pregnant women, most of whom are young and unmarried and enroll during their second or third trimester. It has had impressive success in improving health, such as reductions in smoking during pregnancy, maternal hypertension, rates of pre-term and low-birth-weight births, and children’s emergency room visits, and also appears to affect the timing and number of subsequent births. Children whose mothers participate in the program see modest improvements in their school readiness.

Preschool programs emphasize improving children’s academic skills more than their health or home environments. Preschool attendance is one of the strongest predictors of school readiness. Income and racial gaps in academic and social competencies are evident as early as age 4 or 5, and compensating for what these more disadvantaged children do not learn at home is a highly effective strategy. These programs have an impact not only on school readiness but also on later outcomes. For example, a child who is school-ready is almost twice as likely to acquire core competencies by age 11 and, having achieved those competencies, to go on to graduate from high school and be successful as an adult. Some evidence also suggests that school-ready children go on to do better in the labor market even if their early experience does not affect their test scores in later grades, perhaps because it affects their social skills, their self-discipline, or their sense of control over their lives.

Middle childhood. The period from school entry, around age 5, until the end of elementary school, around age 11, is often overlooked by policymakers. Yet it is the time when most children master the basics—reading and math—as well as learn to navigate the world outside the home.

Scores on math and reading, especially the latter, have improved little in recent decades. Gaps by race have narrowed only modestly, whereas gaps by income have widened dramatically. These academic skills will be much more important in the future as the economy sheds routine production and administrative jobs and demands workers with higher-level critical thinking abilities. For these reasons, almost everyone agrees that the education system needs to be transformed. School reform could include more resources, more accountability for results, more effective teachers in the classroom, smaller class sizes, more effective curricula, longer school years or days, and more competition and choice via vouchers or charter schools.

However, most experts do not believe that more resources by themselves will have much impact unless they are targeted in effective ways. They also agree that accountability is important but must be combined with providing schools the capacity to do better, that teachers are critical but that it is hard to identify good teachers in advance, and that class size (holding teacher quality constant) matters but is a comparatively expensive intervention. Few curriculum reforms have been well evaluated or have demonstrated big effects. Finally, some charter schools and voucher experiments have produced positive results, but charters as a whole do not do a better job than the public schools.

In short, there is no simple solution. Most likely a combination of these or other reforms will be needed to improve children’s competencies in the middle years. Indeed, more-holistic approaches or “whole-school reforms,” such as Success For All, that involve simultaneously changing teacher training, curricula, testing, and the organization of learning have had some success. It is also encouraging that national benchmarks in the form of the Common Core State Standards have now been endorsed by 45 states. The Obama administration has pressed for tracking children’s progress in school, for rewarding teachers based in part on how much children learn, and for more innovation through charter schools. The nation is also going to need new experiments in the uses of technology and online learning. In the meantime, there are numerous more-limited efforts that have had some success at improving reading or math and that could be expanded to more schools.

It is not only academic skills that matter. Our data show that many children lack the behavioral skills that recent research indicates are important for later success. Interventions designed to improve children’s social-emotional competencies in the elementary-school years have produced promising results and need to be part of the solution. Social-emotional learning has five core elements: self-awareness, self-management, social awareness, relationship skills, and responsible decisionmaking. Learning to navigate these areas positively affects children’s behavior, reduces emotional distress, and can indirectly affect academic outcomes.

Finally, our analyses suggest that most children are reasonably healthy at age 11. For those who are not, access to health care is obviously important. It is likely that between Medicaid, the State Children’s Health Insurance Program, and the Patient Protection and Affordable Care Act, most children are (or will be) covered by public or private health insurance. But important exceptions may include illegal immigrants, children who live in rural areas, or children whose parents fail to bring them in for care. In the latter context, a recent experiment in New York City, the Family Rewards program, in which parents were offered a reward for making sure their children had insurance and regular checkups, found that there were only modest effects on the receipt of care, which is already quite high in the coverage area. But making sure that children get dental checkups, immunizations, and other preventive care is worth pursuing.

Adolescence. Our data show that just 57% of children graduate from high school with decent grades and having avoided teen parenthood and criminal conviction. Dropping out of high school, not learning very much during the high-school years, and engaging in risky behaviors are all barriers to later success. There has been a rise in incarceration during the past 30 years, driven by patterns among young men, particularly African American men. In contrast, and further reflecting the diverging trends for young men and women, teen births have declined to historically low levels—a consequence of historically low pregnancy rates, although rates remain high by international standards.

The teenage years have always been fraught with difficulty, but these data should be a wakeup call for parents, schools, community leaders, elected officials, and young people themselves. By the time they are teens, young people should begin to take some responsibility for their futures, and they need to be encouraged by parents and others to do so.

At the same time, school reforms can make a difference. In many studies, high-school dropouts report high levels of disengagement and lack of motivation. Education leaders and policymakers are attempting to respond to these issues of engagement with new types of high schools. One example, from New York City, is the conversion of a number of failing public high schools to small, academically nonselective, four-year high schools that emphasize innovation and strong relationships between students and faculty. The schools, dubbed “small schools of choice,” are intended to serve as alternatives to the neighborhood schools located in historically disadvantaged communities. Admission is determined by a lottery system after students across New York City select and rank the high schools they would like to attend.

The random assignment of students to these schools enabled researchers to study their effect on various student outcomes. The results indicate that enrollment in these schools had a substantial impact on student achievement. Throughout high school, these students earned more credits, failed fewer courses, attended school more dependably, and were more likely to be on track to graduate. They earned more New York State Regents diplomas, had a higher proportion of college-ready English exam scores, and had graduation rates nearly 15% higher than schools examined as controls.

Young adulthood. The transition to adulthood is increasingly lengthy, with more young adults living at home, fewer of them marrying, and more of them continuing their educations well into their 20s or beyond. This makes defining success during this period difficult. We look at what proportion of young adults are living independently and have graduated from college or are in a household with an income above 250% of poverty by the end of their 20s.

One very important barrier to college is inadequate preparation. As our data show, individuals who have graduated from high school with decent grades and no involvement in risky behavior are twice as likely to complete college and have a 71% chance of achieving success by the end of their 20s, whereas those who do not have only a 45% chance. But academic preparation is not all that matters. Research has shown that even among students with equal academic achievements, socioeconomic status has a large effect on who finishes college: 74% of high scorers who grew up in upper-income families completed college, compared with 29% of those who grew up in low-income families. Moreover, these income gaps in college attendance have widened in recent years.

There are a multitude of federal and state programs, tax credits, and loans that subsidize college attendance. One of the most important for closing gaps between more and less advantaged children is the Pell Grant program, because it is targeted to lower-income families (those with incomes below about $40,000 a year). The federal government spends about $36 billion on the program annually. The evidence on the effects of the grants is somewhat mixed. In part, this is because even if financial incentives matter (and they appear to, according to the best research), the process of applying for grants is daunting for lower-income families. Ongoing reforms to simplify the process should help. It has been estimated that for each additional $1,000 in subsidies, the chance of college enrollment increases by about 4 percentage points.

Whether enrollment leads to graduation is another matter. Although enrollment rates in postsecondary institutions, including community colleges, have shot up, graduation rates have increased very little. Efforts to improve retention and graduation rates by coupling tuition assistance with more personalized attention and services have had only modest effects to date.

Adulthood. Once individuals have completed schooling, left home, and entered the work force, the most important determinant of their success is the labor market. For those who reach adulthood without the academic and social skills to enter the middle class on their own, government policies should be linked to personal responsibility and opportunity-enhancing behaviors. This includes education subsidies tied to academic performance; income assistance that encourages and rewards work, such as the Earned Income Tax Credit and child-care subsidies; career and technical education tied to growing sectors of the economy; and more apprenticeship and on-the-job training. Although important for all people during this aftermath of the Great Recession, such programs may be especially important for less-well-educated individuals, who have suffered not just a loss of employment opportunities but a decline in their earnings.

Putting the full responsibility on government to close all of these gaps across all of the life stages is, of course, unreasonable. But so is a heroic assumption that everyone can be a Horatio Alger with no help from society. By drawing on the lessons of science, government can put in place effective programs to boost the success of everyone striving to reach the middle-class dream, while also saving taxpayers money over the long run.

The 21st-Century Land Rush

The International Food Policy Research Institute (IFPRI) recently reported that nearly three dozen countries designated as “alarming” or “serious” on its Global Hunger Index scale are leasing vast swaths of farmland to international investors. IFPRI revealed that these foreign investors are cultivating crops and then immediately spiriting them out of these hunger-riven nations. Seven countries with high hunger levels—Cambodia, Ethiopia, Indonesia, Laos, Liberia, the Philippines, and Sierra Leone—have each agreed to deals that collectively constitute more than 10% of their total agricultural area (the figure approaches 20% in Sierra Leone). Oxfam just announced a similar finding: Two-thirds of large-scale land acquisitions have occurred in countries with “a serious hunger problem.”

Why are the world’s most food-insecure nations ceding precious food resources to outsiders? The answer requires a closer look at these land deals, which IFPRI has described as a new phase of the world food crisis.

According to Oxfam, since 2001 foreigners have acquired nearly 230 million hectares of farmland (the size of Western Europe), with most of this land being obtained since 2008. According to the World Bank, 60 million hectares’ worth of deals were announced in 2009 alone. The amount of relinquished land in poor nations, says Oxfam, equates to an area the size of London sold off every six days.

Asian (particularly Chinese, South Korean, and Indian), European, and Gulf governments and corporations, in pursuit of food and energy resources or simply seeking profits, are snapping up precious arable land in sub-Saharan Africa, Southeast Asia, Latin America, and the former Soviet Union. Most of these target areas are food-insecure (some are dependent on international food aid), impoverished, and corruption-prone, and few of their citizens enjoy robust land rights.

This is not to say, however, that large-scale land acquisitions are strictly a matter of wealthy investors preying on the developing world’s destitute. Africans and Asians are both investing intraregionally, and Australia and New Zealand have emerged as attractive investment destinations. The largest potential land deal announced to this point—a staggering 6 million hectares to grow soy and corn in Mozambique—has been pursued by Brazil.

This global land rush is not new. Several centuries ago, European colonialists and U.S. fruit corporations gobbled up enormous expanses of global farmland. However, several characteristics set today’s investments apart from their predecessors. One, suggested by the Brazil-Mozambique example, is sheer scale. Ten deals announced since 2007 involve at least 100,000 hectares of land, and there are five deals that constitute at least a million hectares. A study released this past summer by the United Nations and International Institute for Environment and Development (IIED) concludes that 2.5 million hectares have been acquired in just five countries (Ghana, Ethiopia, Madagascar, Mali, and Sudan), including almost 1.5 million hectares in Sudan alone. To get a sense of the magnitude, consider that the typical developing-world farmer has a plot of just two hectares.

Another difference from the past is product type. In previous centuries, the focus was on cash crops such as tea and bananas. Today, the emphasis is on grains and potential biofuels such as wheat, rice, sorghum, soy, corn, palm oil, and sugarcane. These are staples that meet the food and energy needs of millions.

Problematic deals

Given the need for greater investment in developing-world agriculture—a low-yield, inefficient sector whose struggles helped precipitate the 2007–2008 world food crisis—one may point to these land deals as an encouraging development that will boost agricultural productivity and benefit the local economy. Unfortunately, they are deeply problematic.

Generally speaking, they have not resulted in benefits for local communities, because most of the deals hire very few local laborers, transfer few agricultural technologies to local communities, and, most critically, sell few harvests to local markets. According to a study released by European researchers earlier this year, about two-thirds of foreign land investors operating in developing nations intend to export all of their production. Many investors don’t even bother to farm their land, instead preferring to sit on it for speculative purposes. A 2011 World Bank study found that farming has begun on only about 20% of publicly announced deals.

Projects that have been implemented pose considerable environmental threats. Land investors favor large-scale industrial agriculture, which emphasizes fossil fuel–based technologies, pesticides, and fertilizers. They also embrace deep plowing and heavy water use, which degrade land and tax natural resources. Deforestation is a major concern as well. Indigenous communities in Indonesia have lodged a formal complaint with the World Bank that accuses a palm oil company of destroying their forests. Yet the problem extends far beyond Indonesia. My new edited volume, The Global Farms Race, features research led by University of Georgia professor Laura German. She reveals that in both Southeast Asia and sub-Saharan Africa, forest conversions to oil palm production (one of the major crops figuring in land deals) have resulted in 100% deforestation rates. They have also caused extensive damage to indigenous animal species, including the elimination of 60% of a bird species in Malaysia.

Additionally, farmland investments imperil the basic needs of local communities by appropriating land they have long accessed for food, water, and medicine. In Ethiopia, an Indian agribusiness firm is growing food for export on land that previously cultivated teff, Ethiopia’s staple grain. Even worse, researchers and journalists have documented numerous cases of people suffering outright displacement. Land rights organizations allege that a Thai sugar company’s acquisition of 20,000 hectares in Cambodia uprooted hundreds of locals. According to Human Rights Watch, a resettlement program has “forcibly displaced” tens of thousands of Ethiopians in Gambella state, where more than 40% of the land has been set aside for investment. And in Uganda, a British firm’s seemingly benign project—the planting of new forest land—has displaced 20,000 people.

Some deals have sparked community unrest. In 2009, an unsuccessful attempt by a South Korean corporation to acquire 1.3 million hectares of farmland in Madagascar set off widespread national protests and helped oust the government that had championed the deal. More recently, in 2011, a Ugandan mob, angry about an Indian firm’s plan to clear rainforest for sugarcane production, killed an Indian man. That same year, furious Kenyans told the Guardian newspaper that they had been evicted from the Tana Delta to enable investors to construct a sugar plantation and vowed to respond with “guns and sticks … it will be war.” Another potential flashpoint is Papua, an Indonesian region wracked by a separatist insurgency led by indigenous Papuans. Here, a Saudi firm has acquired more than a million hectares on a Jakarta-controlled estate. With non-Papuans expected to be brought in as laborers, observers fear an exacerbation of strife.

Finally, large-scale land acquisitions could deliver a serious blow to a global food economy already hampered by sluggish harvests, high commodities prices, and recession. Imagine if major food-importing nations, spooked by volatile food markets and fearful of export bans, continue to undertake large-scale food production abroad instead of importing through commodities markets. This could mean lower market demand, reduced supplies, and, in all likelihood, even higher prices for those still dependent on food from market sources. (Food prices have risen by nearly 4% in poor countries since 2011, according to UN estimates.) The most vulnerable would be poor food importers, such as West African nations, with insufficient capital to follow the lead of wealthier importers and invest in farmland overseas.

The intention here is not to paint all land deals with a critical brush; there are undoubtedly success stories. Researchers from IIED have discovered contracts for several deals in Africa that apply international environmental standards and contain explicit investor commitments on employment and training. The World Bank has singled out several projects that generate employment for local smallholders. And New York Times reportage has described how Chinese acquisitions of Russian farmland are generating win-win results for both countries: China produces food and other resources desperately needed back home, yet also sells some of its harvests in local Russian markets, and even hands out free produce to those who happen to stop by its farms. Beijing also supplies scores of its own agricultural laborers, a boon for Russia’s land-rich yet sparsely populated rural regions. Such examples, however, represent exceptions, not norms.

MOST OF THE DEALS HIRE VERY FEW LOCAL LABORERS,
TRANSFER FEW AGRICULTURAL TECHNOLOGIES TO LOCAL COMMUNITIES,
AND, MOST CRITICALLY, SELL FEW HARVESTS TO LOCAL MARKETS
.

Rising stakes

The race for farmland has no end in sight, because the drivers of large-scale land acquisitions promise to persevere in the years ahead. These include volatile commodities markets, extreme weather events, disappearing natural resources, population growth, and skyrocketing demand for food and biofuels.

Additionally, investors have powerful incentives to push forward. With food security under constant threat, food-importing governments must obtain food for their populations, and by producing it overseas, they bypass the perils of unpredictable markets. Meanwhile, private-sector financiers have compelling commercial reasons to continue to hoard land overseas, given high global demand for food (and biofuels) and the dwindling global supply of land and water resources needed for its production.

As long as the land deals continue, the threats outlined above will remain and could grow more acute as investments continue to proliferate. Imagine the long-term environmental implications of large-scale farming projects across central Africa, South America, and Southeast Asia, regions that not only host many land deals but also are home to most of the world’s shrinking rainforests. Consider how the intensification of large-scale plantation agriculture could erode smallholder agriculture and snuff out local farming skills. And contemplate the possibility of long-time regional food export hubs, such as Southeast Asia’s rice production areas, becoming net importers, as outside investors devour local cropland and send their harvests back home, leaving the region with insufficient land to maintain its own export regime. Raul Montemayor, who runs a small farmers’ organization in the Philippines, warns in The Global Farms Race that large-scale land acquisitions, coupled with Asia’s rice demand, population growth, and expected climate change effects along the Mekong River, could well spark such a shift.

What can be done?

Because the chief motivations and incentives remain so strong, little can be done to halt large-scale land acquisitions. The international community has no ability (much less the right) to prevail on investors, most of them wealthy and powerful, to stop their activities. Therefore, Oxfam’s recent demand that the World Bank suspend its financial support for large-scale acquisitions is misguided.

Additionally, unlike in the past, host governments enthusiastically support these investments, because of the bridges, ports, and roads that investors promise to incorporate into their projects (large revenue boosts for host government coffers certainly don’t hurt either). Host officials have offered a range of inducements to foreign investors, from tax holidays to private security forces. Prevailing on host governments to stop signing off on such land projects, therefore, is unrealistic as well.

The most practical strategy is to accept the existence of the land deals and encourage policy changes that blunt their harmful effects. National governments seeking agricultural investments should tighten regulations governing foreign acquisitions of land, and they should make more vigorous efforts to limit the damage these deals cause to food security, livelihoods, and the environment. These governments should also establish greater protections for vulnerable smallholders. One example is to provide legal assistance to small farmers to ensure that their interests are better protected in contract negotiations. Another is facilitating the ability of citizens to secure land titles, which would make them less vulnerable to land predations. The charity Concern Worldwide recently helped 6,000 Tanzanians obtain such documents in a country where only 10% of the population holds them.

Several host governments are already taking such measures. Argentina and Brazil are passing laws to limit foreign land ownership. Cambodia and Laos have even declared freezes on new land concessions. Admittedly, however, there is no guarantee that these policies will achieve their intended goals. Additionally, they have yet to be established in most of the popular African targets of land acquisitions, such as Ethiopia, Sudan, and Kenya.

The international community also has an important role to play. Normative measures, such as the implementation of international codes of conduct to ensure equitable and sustainable land deals, have little utility, given that such efforts are unenforceable and unlikely to be heeded voluntarily. Instead, international nongovernmental organizations and the media should showcase success stories, such as those few deals that actually deliver on investor promises and those governments that have acted to mitigate the deleterious aspects of land deals. The international community should also seek to shame the most egregious investors by producing extensive reportage on the most damaging deals.

Even while establishing measures to minimize the harmful effects of these investments today, additional efforts should take a more long-term view and strengthen the alternatives to large-scale land acquisitions, so that nations have more-innocuous ways of securing food sources when commodities markets are too volatile. The private-sector firm Syngenta is developing genetically engineered drought-resistant wheat and corn crops. This innovation could eventually provide parched Gulf nations with an indigenous solution to food insecurity that makes forays abroad unnecessary. And in another major development, after years of political resistance, Southeast Asian nations have finally agreed to form a regional rice pool. If such plans are consummated, these countries may have less of an incentive to turn to outside investors to enhance food production.

Finally, efforts should be made to ramp up international investments in agriculture that focus less on large-scale production and more on small-scale, targeted initiatives. Investors can finance irrigation development, higher-yield farming technologies, and training courses that enhance local communities’ agricultural skills.

Natural resource doomsday?

Increasingly, natural resource experts warn that the world is entering a Hobbesian period of resource scarcity in which nations must fight for the precious resources that remain. This era represents, to quote the title of Michael T. Klare’s newest book, “the race for what’s left.” Lester Brown’s new book, Full Planet, Empty Plates, tells a similarly scary story of dwindling food stocks, crop yields, and water and land resources colliding with rapid population growth and soaring food demand. “We are moving into a new food era,” he recently told the Guardian, “one in which it is every country for itself.”

Some may dismiss this talk as needless alarmism, but the story of large-scale land acquisitions seems to foreshadow this desperate, no-holds-barred resource scramble. Food-needy nations are taking matters into their own hands by eschewing food markets and growing their own food abroad. In most host nations, smallholder communities can only watch helplessly as these outsiders voraciously consume their precious land and water resources. Meanwhile, less capital-rich food importers, or those simply unwilling to venture abroad to obtain food, must make do with volatile and expensive food markets, and worry about the ramifications for their restive and hungry populations.

The takeaway? The global farms race is freighted with significance—for its participants, its spectators, and even those who, at their own peril, pay it little mind.

Michael Kugelman () is the senior program associate for South and Southeast Asia at the Woodrow Wilson International Center for Scholars in Washington, DC. He is the lead editor of The Global Farms Race: Land Grabs, Agricultural Investment, and the Scramble for Food Security. Follow him on Twitter .

Real Numbers: A Perpetual Imbalance?

Federal Funding of Physical Sciences and Engineering Research

A major theme of the National Academies’ 2007 report, Rising Above the Gathering Storm, and a major premise of the 2007 America COMPETES Act and its 2010 reauthorization was that the nation urgently needed to boost funding of research in the physical sciences and engineering (PS&E) after a decade-long decline precipitated by the end of the Cold War.

The COMPETES Act authorized a doubling of research funding over seven years (later extended to eleven years in the 2010 reauthorization) through the National Science Foundation, National Institute of Standards and Technology (NIST), and Department of Energy (DOE) Office of Science on the reasonable presumption that much of the increase would go into PS&E fields. Interestingly, on the same premise, Rising Above the Gathering Storm singled out only Department of Defense (DOD) research as meriting increased support. It was perhaps to be expected that Congress would focus on civilian agency rather than military service budgets to rectify the decade-long neglect of most PS&E fields, but shifting the attention had important consequences.

Although the COMPETES reauthorization has a year to run, it is not premature to assess progress in righting the “imbalance” in the federal research portfolio. The Congressional Research Service estimated that if current trends continue, the desired doubling will not take place in 7 years or 11 years or even 15 years. A review of actual funding from 2001 to 2011 (using data available from the National Science Foundation’s federal funds survey) reveals that funding of the physical sciences is virtually flat and funding of engineering has declined.

Stephen A. Merrill () is director of the National Academies’ Board on Science, Technology, and Economic Policy.

The dark years

Between FY 1993 and FY 1999 federal support of physics, chemical engineering, geological sciences, and electrical and mechanical engineering declined by more than 20% in real terms, while funding of the biological and medical sciences increased by more than 20% in the same period. Computer science funding was the major exception to the PS&E pattern, rising more than 60%.

Shifts in Federal Research Support, FY 1993-1999

Expectations unrealized

Funding of physical science research increased only slightly (0.3%) from 2001 to 2011, and engineering research has actually declined by 4.3%, while overall research expenditures rose 8.1%, and life sciences increased 7.0%.

These figures exclude the two years of American Recovery and Reinvestment Act (ARRA) research funding because the National Science Foundation decided not to add to agencies’ reporting burden for “stimulus” expenditures by asking them to disaggregate them by research field.

Federal funding of broad felds of research, FY 2001-2011 (not including ARRA)

Stimulus funding not sufficient

But it turns out that three-quarters of the $13.1-billion (in 2005 dollars) ARRA boost to research spending in FY 2009 and 2010 went to the National Institutes of Health, primarily for research in the biological and medical sciences.

American Recovery and Reinvestment Act (ARRA) research funding, FY 2009 & FY 2010
(2005 constant $ millions)

Engineering particularly hard hit

Less widely recognized is the continued steep decline (down 26% from FY 2001 to FY 2010) of engineering research support by DOD, which still accounts for one-third of all federal investment in engineering. DOD accounts for 28% of federal math/ computer science funding and 14% of federal physical science funding.

DOD Mathematics/CS, Physical Sciences & Engineering research funding, FY 2001-2010

RIP: The Basic/Applied Research Dichotomy

U.S. science policy since World War II has in large measure been driven by Vannevar Bush’s famous paper Science—The Endless Frontier. Bush’s separation of research into “basic” and “applied” domains has been enshrined in much of U.S. science and technology policy over the past seven decades, and this false dichotomy has become a barrier to the development of a coherent national innovation policy. Much of the debate centers on the appropriate federal role in innovation. Bush argued successfully that funding basic research was a necessary role for government, with the implication that applied research should be left to the auspices of markets. However, the original distinction does not reflect what actually happens in research, and its narrow focus on the stated goals of an individual research project prevents us from taking a more productive holistic view of the research enterprise.

By examining the evolution of the famous linear model of innovation, which holds that scientific research precedes technological innovation, and the problematic description of engineering as “applied science,” we seek to challenge the existing dichotomies between basic and applied research and between science and engineering. To illustrate our alternative view of the research enterprise, we will follow the path of knowledge development through a series of Nobel Prizes in Physics over several decades.

This mini-history reveals how knowledge grows through a richly interwoven system of scientific and technological research in which there is no clear hierarchy of importance and no straightforward linear trajectory. Accepting this reality has profound implications for the design of research institutions, the allocation of resources, and the national policies that guide research. This in turn can open the door to game-changing discoveries and inventions and put the nation on the path to a more sustainable science and technology ecosystem.

History of an idea

Although some observers cite Vannevar Bush as the source of the linear model of innovation, the concept actually has deep roots in long-held cultural assumptions that give priority to the work of the head over the work of the hand and thus to the creation of scientific knowledge over technical expertise. If one puts this assumption aside, it opens up a new way of understanding the entire innovation process. We will focus our attention on how it affects our understanding of research.

The question of whether understanding always precedes invention has long been a troubling one. For example, it is widely accepted that many technologies reached relatively advanced stages of development before detailed scientific explanations about how the technologies worked emerged. In one of the most famous examples, James Watt invented his steam engine before the laws of thermodynamics were postulated. In fact, the science of thermodynamics owes a great deal to the steam engine. This and other examples should make it clear that assumptions about what has been called basic and applied research do not accurately describe what actually happens in research.

In 1997, Donald Stokes’s book Pasteur’s Quadrant: Basic Science and Technological Innovation was published posthumously. In this work, Stokes argued that scientific efforts were best carried out in what he termed “Pasteur’s Quadrant,” where researchers are motivated simultaneously by expanding understanding and increasing our abilities (technological, including medicine) to improve the world. Stokes’s primary contribution was in expanding the linear model into a two-dimensional plane that sought to integrate the idea of the unsullied quest for knowledge with the desire to solve a practical problem.

Stokes’s model comprises three quadrants, each exemplified by a historical figure in science and technology. The pure basic research quadrant exemplified by Niels Bohr represents the traditional view of scientific research as being inspired primarily by a desire to extend fundamental understanding. The pure applied research quadrant is exemplified in Edison, who represents the classical inventor, driven to solve a practical problem. Louis Pasteur’s quadrant is a perfect mix of the two, inventor and scientist in one, expanding knowledge in the pursuit of practical problems. Stokes described this final quadrant as “use-inspired basic research.” The fourth quadrant is not fully described in Stokes’ framework.

The publication of Stokes’s book excited many in the science policy and academic communities, who believed it would free us from the blinders of the linear model. A blurb on the back of the book quotes U.S. Congressman George E. Brown Jr.: “Stokes’s analysis will, one hopes, finally lay to rest the unhelpful separation between ‘basic’ and ‘applied’ research that has misinformed science policy for decades.” However, it has become clear that although Stokes’s analysis cleared the ground for future research, it did not go far enough, nor did his work result in sufficient change in how policymakers discuss and structure research. Whereas Stokes notes how “often technology is the inspiration of science rather than the other way around,” his revised dynamic model does not recognize the full complexity of innovation, preferring to keep science and technology in separate worlds that mix only in the shared agora of “use-inspired basic research.” It is also significant that Stokes’s framework preserves the language of the linear model in the continued use of the terms basic and applied as descriptors of research.

We see a need to jettison this conception of research in order to understand the complex interplay among the forces of innovation. We propose a more dynamic model in which radical innovation often arises only through the integration of science and technology.

Invention and discovery

A critical liability of the basic/applied categorization is that it is based on the motivation of the individual researcher at the time of the work. The efficacy and effectiveness of the research endeavor cannot be fully appreciated in the limited time frame captured by a singular attention to the motivations of the researchers in question. Admittedly, motivations are important. Aiming to find a cure for cancer or advance the frontiers of communications can be a powerful incentive, stimulating groundbreaking research. However, motivations are only one aspect of the research process. To more completely capture the full arc of research, it is important to consider a broader time scale than that implied by just considering the initial research motivations. Expanding the focus from research motivations to also include questions of how the research is taken up in the world and how it is connected to other science and technology allows us to escape the basic/applied dichotomy. The future-oriented aspects of research are as important as the initial motivation. Considering the implications of research in the long term requires an emphasis on visionary future technologies, taking into account the well-being of society, and not being content with a porous dichotomy between basic and applied research.

We propose using the terms “invention” and “discovery” to describe the twin channels of research practice. For us, invention is the “accumulation and creation of knowledge that results in a new tool, device, or process that accomplishes a specific purpose.” Discovery is the “creation of new knowledge and facts about the world.” Considering the phases of invention and discovery along with research motivations and institutional settings enables a much more holistic and long-term view of the research process. This allows us to examine the ways in which research generates innovation and leads to further research in a virtuous cycle.

Innovation is a complex, nonlinear process. Still, straightforward and sufficiently realized representations such as Stokes’s Pasteur’s quadrant are useful as analytical aids. We propose the model of the discovery-invention cycle, which will serve to illustrate the interconnectedness of the processes of invention and discovery, and the need for consideration of research effectiveness over longer time frames than is currently the case. Such a model allows for a more reliable consideration of innovation through time. The model could also aid in discerning possible bottlenecks in the functioning of the cycle of innovation, indicating possible avenues for policy intervention.

A family of Nobel Prizes

To illustrate this idea, consider Figure 1 below, in which we trace the evolution of the current information and communication age. What can be said about the research that has enabled the recent explosion of information and communication technologies? How does our model enable a deeper understanding of the multiplicity of research directions that have shaped the current information era? To fully answer this question, it is necessary to examine research snapshots over time, paying attention to the development of knowledge and the twin processes of invention and discovery, tracing their interconnections through time. To our mind, the clearest place for selecting snapshots that illustrate the evolution of invention and discovery that enables the information age is the Nobel Prize awards.

We have thus examined the Nobel Prizes in Physics from 1956, 1964, 1985, 1998, 2000, and 2009, which were all related to information technologies. We describe these kinds of clearly intersecting Nobels as a family of prizes in that they are all closely related. Similar families can be found in areas such as nuclear magnetic resonance and imaging.

The birth of the current information age can be traced to the invention of the transistor. This work was recognized with the 1956 Physics Nobel Prize awarded jointly to William Shockley, John Bardeen, and Walter Brattain “for their researches on semiconductors and their discovery of the transistor effect.” Building on early work on the effect of electric fields on metal semiconductor junctions, the interdisciplinary Bell Labs team built a working bipolar-contact transistor and clearly demonstrated (discovered) the transistor effect. This work and successive refinements enabled a class of devices that successfully replaced electromechanical switches, allowing for successive generations of smaller, more efficient, and more intricate circuits. Although the Nobel was awarded for the discovery of the transistor effect, the team of Shockley, Bardeen, and Brattain had to invent the bipolar-contact transistor to demonstrate it. Their work was thus of a dual nature, encompassing both discovery and invention. The discovery of the transistor effect catalyzed a whole body of further research into semiconductor physics, increasing knowledge about this extremely important phenomenon. The invention of the bipolar contact transistor led to a new class of devices that effectively replaced vacuum tubes and catalyzed further research into new kinds of semiconductor devices. The 1956 Nobel is therefore exemplary of a particular kind of knowledge-making that affects both later discoveries and later inventions. We call this kind of research radical innovation. The 1956 prize is situated at the intersection of invention and discovery (see Figure 1), and it is from this prize that we begin to trace the innovation cycle for the prize family that describes critical moments in the information age.

FIGURE 1

The innovation cycle in information and communication technologies (dates of events are in red).

The second prize in this family is the 1964 Nobel Prize, which was awarded jointly to Charles Townes and the other half to both Nicolay Basov and Aleksandr Prokhorov. Most global communications traffic is carried by transcontinental fiber optic networks, which use light as the signal carrier. Townes’s work on the stimulated emission of microwave radiation earned him his half of the Nobel. This experimental work showed that it was possible to build amplifier oscillators with low noise characteristics capable of the spontaneous emission of microwaves with almost perfect amplification. The maser (microwave amplification by the stimulated emission of radiation effect) was observed in his experiments. Later, Basov and Prokhorov, along with Townes, extended the maser effect to consideration of its application in the visible spectrum, and thus the laser was invented. Laser light allows for the transmission of very high-energy pulses of light at very high frequencies and is crucial for modern high-speed communication systems. This Nobel acknowledges critical work that was also simultaneously discovery (the maser effect) and invention (the maser and the laser), both central to the rise of the information and communication age. Thus, the 1964 Nobel is also situated at the intersection of invention and discovery. The work on lasers built directly on previous work by Einstein, but practical and operational masers and lasers were enabled by advancements in electronic amplifiers made possible by the solid-state electronics revolution, which began with the invention of the transistor.

Although scientists and engineers conducted a great deal of foundational work on the science of information technology in the 1960s, the next wave of Nobel recognition for this research did not come until the 1980s. Advancements in the semiconductor industry led to the development of new kinds of devices such as the metal oxide silicon field effect transistor (MOSFET). The two-dimensional nature of the conducting layer of the MOSFET provided a convenient avenue to study electrical conduction in reduced dimensions. Klaus von Klitzing discovered that under certain conditions, voltage across a current-carrying wire increased in uniform steps. Von Klitzing received the 1985 Nobel Prize for what is known as the quantized Hall effect. This work belongs in the discovery category, although it did have important useful applications.

The 2000 Nobel Prize was awarded jointly to Zhores Alferov and Herbert Kroemer for “developing semiconductor heterostructures” and to Jack Kilby for “his part in the invention of the integrated circuit.” Both of these achievements can be classified primarily as inventions, and both built on work done by Shockley et al. This research enabled a new class of semiconductor device that could be used in high-speed circuits and optoelectronics. Alferov and Kroemer showed that creating a double junction with a thin layer of semiconductors would allow for much higher concentrations of holes and electrons, enabling faster switching speeds and allowing for laser operation at practical temperatures. Their invention produced tangible improvements in lasers and light-emitting diodes. It was the work on heterostructures that enabled the modern room-temperature lasers used in fiber optic communication systems. Alferov and Kroemer’s work on heterostructures also led to the discovery of a new form of matter, as discussed below.

Jack Kilby’s work on integrated circuits at Texas Instruments earned him his half of the Nobel for showing that entire circuits could be realized with semiconductor substrates. Shockley, Bardeen, and Brattain had invented semiconductor-based transistors, but these were discrete components and were used in circuits with components made from other materials. The genius of Kilby’s work was in realizing that semiconductors could be arranged in such a way that the entire circuit, not just the transistor, could be realized on a chip. This invention of a process of building entire circuits out of semiconductors allowed for economies of scale, bringing down the cost of circuits. Further research into process technologies allowed escalating progress on the shrinking of these circuits, so that in a few short years, chips containing billions of transistors were possible.

Alferov and Kroemer’s work was also valuable to Horst Stormer and his collaborators, who combined it with advancements in crystal growth techniques to produce two-dimensional electron layers with mobility orders of magnitude greater than in silicon MOSFETs. Stormer and Daniel Tsui then began exploring some observed unusual behavior that occurred in two-dimensional electrical conduction. They discovered a new kind of particle that appeared to have only one-third the charge of the previously thought-indivisible electron. Robert Laughlin then showed through calculations that what they had observed was a new form of quantum liquid where interactions between billions of electrons in the quantum liquid led to swirls in the liquid behaving like particles with a fractional electron charge. This phenomenon is clearly a new discovery, but it was enabled by previous inventions and resulted in important practical applications such as the high-frequency transistors used in cell phones. For their work, Laughlin, Stormer, and Tsui were awarded the 1998 Nobel Prize in Physics, an achieve ment situated firmly in the discovery category.

The 2009 Nobel was awarded to Charles Kao for “groundbreaking achievements concerning the transmission of light in fibers for optical communication” and to Willard Boyle and George Smith for “the invention of the imaging semiconductor circuit—the CCD.” Both of these achievements were directly influenced by previous inventions and discoveries in this area. Kao was primarily concerned building a workable waveguide for light for use in communications systems. His inquiries led to astonishing process improvements in glass production, as he predicted that glass fibers of a certain purity would allow long-distance laser light communication. Of course, the work on heterostructures that allowed for room-temperature lasers was critical to assembling the technologies of fiber communication. Kao, however, not only created new processes for measuring the purity of glass but also actively encouraged various manufacturers to improve their processes in this respect. Working directly in industry, Kao’s work built on the work by Alferov and Kromer, enabling the physical infrastructure of the information age. Boyle and Smith continued the tradition of Bell Labs inquiry. Adding a brilliant twist to the work that Shockley et al. had done on the transistor, they designed and invented the charge-coupled device (CCD), a semiconductor circuit that enabled digital imagery and video. Kao’s work was clearly aimed at discovering the ideal conditions for the propagation of light in fibers of glass, but he also went further in shepherding the invention and development of the new fiber optic devices.

These six Nobel Prizes highlight the multiple kinds of knowledge that play into the innovations that have enabled the current information and communications age. From the discovery of the transistor effect, which relied on the invention of the bipolar junction transistor and led to all the marvelous processors and chips in everything from computers to cars, to the invention of the integrated circuit, which made the power of modern computers possible while shrinking their cost and increasing accessibility. The invention of fiber optics built on previous work on heterostructures and made the physical infrastructure and speed of the global communications networks possible. In fact, the desire to improve the electrical conductivity of heterostructures led to the unexpected discovery of fractional quantization in two-dimensional systems and a new form of quantum fluid. Each of these could probably be classified as “basic” or “applied” research, but that classification obscures the complexity and multiple nature of the research described above and does not help remove the prejudices of many against what is now labeled as “applied research.” Thinking in terms of invention and discovery through time helps reconstruct the many pathways that research travels along in the creation of radical innovations.

In our model, the discovery-invention cycle can be traversed in both directions, and research knowledge is seen as an integrated whole that mutates over time (as it traverses the cycle). The bidirectionality of the cycle reflects the reality that inventions are not always the product of discovery but can also be the product of other inventions. Simultaneously, important discoveries can arise from new inventions. Observing the cycle of research over time is essential to understanding how progress occurs.

Seeing with fresh eyes

The switch from a basic/applied nomenclature to discovery-invention is not a mere semantic refinement. It enables us to see the entire research enterprise in a new way.

First, it eliminates the tendency to see research proceeding on two fundamentally different and separate tracks. All types of research interact in complex and often surprising ways. To capitalize on these opportunities, we must be willing to see research holistically. Also, by introducing new language, we hope to escape the cognitive trap of thinking about research solely in terms of the researcher’s initial motivations. All results must be understood in their larger context.

Second, adopting a long time frame is essential to attaining a full understanding of the path of research. The network of interactions traced in the Nobel Prizes discussed above becomes clear only when one takes into account a 50-year history. This extended view is important to understanding the development of both novel science and novel technologies.

Third, the discovery-invention cycle could be useful in identifying problematic bottlenecks in research. Once we recognize the complex interrelationship of discovery and invention, we are more likely to see that problems can occur in many parts of the cycle and that we need to heed the interactions among a variety of institutions and types of research.

Bringing together the notions of research time horizons and bottlenecks, we argue that successful radical innovation arises from knowledge traveling the innovation cycle. If, as argued above, all parts of the innovation process must be adequately encouraged for the cycle to function effectively, then the notion of traveling also emphasizes that we should have deep and sustained communication between scientists and engineers, between theorists and practitioners. Rather than separating researchers according to their motivation, we must strive to bring all forms of research into deeper congress.

This fresh view of the research enterprise can lead us to rethinking the design of research institutions to align with the principles of long time frames, a premium on futuristic ideas, and the encouragement of interaction among different elements of the research ecosystem. This is especially pertinent in the case of the mission-oriented agencies such as the Department of Energy and the National Institutes of Health.

Implications for research policy

The pertinent question is how these insights play out in the messy world of policymaking. First, there is an obvious need to complicate the simple and unhelpful distinction between basic and applied research. The notion of the innovation cycle is a very useful aid in thinking about research holistically. It draws attention to the entirety of research practice and allows one to pose the question of public utility to an entire range of activities.

Second, the nature of the public good, and thus the appropriate role for the federal government, changes. The simple and clear notions of basic and applied were useful in one way: They provided a clear litmus test for limits to federal involvement in the research process. The idea that government funding is necessary to pursue research opportunities that aren’t able to attract private funding is a useful one that has contributed to the long-term well-being and productivity of the nation. But through the lens of the discovery-invention cycle, we can see that it would deny federal funding to some types of research that are essential to long-term progress. We suggest that federal support is most appropriate for research that focuses on long-term projects with clear public utility. The difference here is that such research could have its near-term focus on either new knowledge or new technology.

The public good must be understood over the long term, and the best way to ensure that the research enterprise contributes as much as possible to meeting our national goals is to make funding decisions about discovery and invention research in a long-term holistic context.

Four Technologies and a Conundrum: The Glacial Pace of Energy Innovation

 

The energy sector is the world’s largest market. Reaching billions of users, it accounts for approximately $5 trillion of economic activity and $1.8 trillion of trade annually, representing trillions of dollars of accumulated long-lived capital investments. At the same time, the energy sector is under pressure to change to meet a variety of market and societal pressures. In particular, population and economic growth are driving up total demand; urbanization and industrialization are changing the end uses of energy (toward electrification and transportation); and social concerns about national security and the environment are causing many countries to revamp their primary sources of energy (toward domestic low-carbon options or sustainable options, or both).

Although it is a widely accepted idea that the pace of innovation is accelerating in many fields—especially in consumer electronics—it is not clear whether this phenomenon holds true for the energy industry. Understanding the vintage of the energy industry’s current technological configuration might provide a helpful indicator of its ability to adopt new technologies.

The mix of fuels for the global energy system has been closely tracked for decades by a number of agencies and through a variety of reports with precise records of energy prices, consumption by end-use sectors, and flows of primary energy resources. But the mix of technologies that convert those fuels into useful outcomes has not been quantified, nor are they routinely tracked by reporting agencies, analysts, or observers. Although fuels and end uses are important, it is our contention that conversion technologies also deserve attention.

We have tackled this knowledge gap using the United States as a case study to quantify the energy system’s dependence on discrete energy conversion technologies (by fuel and end-use sector). Although many of the underlying data that we used have been known for a long time, to our knowledge this analysis is the first that comprehensively sorts energy consumption by conversion technologies and quantitatively couples those technologies to the primary fuels and end-use sectors. This assessment leads to some forward-looking recommendations for policymakers and analysts.

Calculating the fuel mix

Four end-use sectors—transportation, industry, commercial, and residential—are responsible for U.S. energy consumption, which in 2010 totaled 98.01 quadrillion British thermal units (BTUs), or quads, according to data from the Energy Information Administration (EIA). By sector, transportation was responsible for 27.51 quads, industry for 30.14 quads, commercial for 18.21 quads, and residential for 22.15 quads. Electricity generation, which accounts for approximately 40% of annual primary energy consumption, is sometimes separated out as its own sector. In this scenario, electricity was responsible for 39.58 quads, transportation for 27.43 quads, industry for 19.98 quads, commercial for 4.18 quads, and residential for 6.84 quads.

These end-use sectors consume various types of fuels, or primary energy sources. In 2010, fossil fuels provided 83% (81.44 quads) of the quads consumed. Among this group, petroleum provided 35.97 quads, natural gas 24.65 quads, and coal 20.82 quads. (Individual totals may not match the cumulative total because of rounding errors in the data.) Nuclear power provided another 8.44 quads. Renewable sources collectively provided 8.05 quads: 4.3 quads from bioenergy, 2.51 quads from hydroelectric, 0.92 quads from wind, 0.21 from geothermal, and 0.11 quads from solar thermal and photovoltaic (PV). From this spectrum, it is apparent that 92% of U.S. energy is produced from only four primary fuels: petroleum, natural gas, coal, and nuclear.

This information is depicted in Figure 1, which is the famous “spaghetti graph,” produced by researchers at Lawrence Livermore National Laboratory that charts the flows of energy consumption from the primary fuels to the end-use sectors. The graph quantitatively reveals many things about the U.S. energy system. For example, it illustrates that the transportation sector is predominately fueled by petroleum, that coal is closely coupled with power generation, and that a majority of energy consumption is released as rejected energy (or waste heat).

FIGURE 1

Energy use by fuels (left side) and end-use sectors (right side)
All numbers in quadrillion BTUs

Source: Energy Information Administration and Lawrence Livermore National Laboratory.

Missing information

Although the EIA’s data reveal a wealth of information about fuel sources and prices, no agency has focused specifically on summarizing annual energy use by conversion technology. However, EIA does publish some data about various conversion devices for the power sector and about various appliances in the commercial and residential sectors. Thus, by carefully reviewing, integrating, and synthesizing data from those and additional EIA reports, we were able to produce a reasonably complete picture of energy conversion technology deployment.

Technology-specific information is known in greatest detail for electricity generation. These technologies include steam turbines, combustion turbines, binary turbines (such as those used in geothermal applications), hydraulic turbines (used with hydroelectric dams), internal combustion engines (used primarily with gasoline and diesel fuel, but also with natural gas), wind turbines, and solar photovoltaic panels.

In other sectors, we had to comb through EIA reports and other data sources and make a number of logical assumptions about technology allocations. A few examples may be illustrative. For the transportation sector, we assumed that all gasoline and ethanol would be consumed in spark-ignition engines, whereas diesel and biodiesel would be consumed in compression-ignition engines. In the industrial sector, we focused on a number of key devices, including boilers, burners, and heaters, that use the most energy. At the commercial and residential scales, widely used appliances comprise such things as space heaters, water heaters, and ovens. Regarding solar energy, although many reports aggregate solar thermal and PV, we separated them, assigning PV (which makes electricity) to the power sector category and end-use solar thermal to the appliance category.

Given its diverse applications, allocating natural gas across the various technologies proved most complicated. Much of natural gas is consumed directly in appliances at the end-use sectors, but it also is consumed in boilers for use with steam turbines, in combustion turbines, and in combustion turbines as an intermediary with steam turbines for generating power in combined cycles. Determining the allocation of natural gas across these sectors required reconciling the information from the power sector data with other data on industrial and residential energy use. In addition, the combustion turbine has two distinct applications: consuming natural gas to generate electrical power (stationary) and consuming jet fuel to generate thrust for transportation (aviation).

TABLE 1

Annual U.S. energy consumption by conversion device in 2010, based on our analysis

The results of our analysis are summarized in Table 1 and illustrated in Figure 2.

As Table 1 and Figure 2 show, four conversion technologies—the steam turbine, combustion turbine, spark-ignition engine, and compression-ignition engine—are responsible for 65.4% of all energy consumption in the United States. The steam turbine (30.59 quads) for power generation and the spark-ignition engine (17.41 quads) for transportation are the two dominant technologies. These are followed by the combustion turbine (9.30 quads), which is primarily for power generation and transportation (aviation), and the compression-ignition engine (6.78 quads), which is primarily for transportation but also has some uses in power generation.

Another 25% of all energy consumption can be attributed to two other (relatively large) categories: industrial devices (primarily boilers, burners, and heaters), and a number of smaller and more varied appliances in the residential and commercial sectors. In addition, nearly 5 quads of energy are consumed as feedstock in making products and materials through a variety of chemical pathways. Rounding out the list are hydroelectric turbines, compressors, and pumps (for pipelines), wind turbines, binary turbines, and solar PV systems.

Within the power and transportation sectors, the dominance of the four leading conversion technologies is even more extensive. Of the 39.6 quads consumed for electricity generation, 90% is provided by the steam turbine (at 75%) and combustion turbine (at 15%) alone. Hydroelectric, wind, solar PV, and binary turbines make up the other 10%. Similarly, transportation is dominated by two technologies: of its 27.4 quads of consumption, 86% is provided by the spark-ignition engine (62%) and compression-ignition engine (24%).

Benchmarking vintage and efficiency

In contrast with the concept of accelerating innovation, the four leading technologies that we identified were, by and large, invented or first demonstrated well over a century ago. The combustion turbine for stationary applications was invented in 1791 by John Barber, the fuel cell in 1842 by Sir William Robert Grove, the spark-ignition engine in 1876 by Nicolaus Otto, the hydroelectric turbine in 1878 by William Armstrong, the solar PV cell in 1883 by Charles Fritts, the steam turbine in 1884 by Charles Parsons, the wind turbine in 1888 by Charles F. Bush, and the compression-engine engine in 1893 by Rudolph Diesel. The most recent is the combustion turbine for aviation applications, invented by Frank Whittle in 1930—still almost a century ago.

In other words, the electricity that powers the advanced hallmarks of the industrialized world (consumer electronics, medical diagnostics, and communications technologies) is generated predominately by late–19th-century technology: steam, hydroelectric, and combustion turbines. Specifically, there was a flurry of innovation over a two-decade span in the late 1870s to early 1890s that yielded most of the conversion devices that power the modern economy.

The modern transition from fossil fuels to renewable energy in the electric industry is primarily to wind and solar power (with biomass and geothermal looming on the horizon). Yet even the wind generator and solar cell technologies date back to the 19th century. Nuclear fission (first demonstrated in 1938, one year after the combustion turbine for aviation applications) is a newer fuel source compared with coal, but it still relies on the same old steam turbine. And the first nuclear power plant was connected to the grid in 1954, almost 60 years ago.

FIGURE 2

Annual U.S. energy use by conversion technology (and fuel and end-use sector) in 2010
All numbers in quadrillion BTUs, based on various sources from the Energy Information Administration

Source: M.E. Webber, UT Austin © 2012.

The United States should invest in R&D for new fundamental energy conversion technologies and concentrate on energy conversion pathways that reduce waste heat.

Solar and wind still meet just a small fraction of annual energy needs. Although improvements in efficiency have greatly increased the performance of these technologies, the fundamental conversions remain the same as in the first solar and wind demonstrations in the 1800s. Thus, even modern fuels such as nuclear still use old technology, and the modern technologies that are available for application in the near term are not as new as commonly thought.

It might therefore be concluded that fundamental innovation in power and transportation technology is slow: Power production does not appear to have an equivalent to the Moore’s Law pace of rapid innovation in the performance and efficiency of microprocessors. This difference has created a strategic problem for planners, who often seem to be betting on breakthroughs to solve the challenges of the energy system. In contrast, the historical record implies that innovation in energy conversion devices is actually quite slow and breakthroughs rare. Thus, the hope that a new technology will come along that suddenly solves the challenge of greenhouse gas emissions seems naive.

The dominance of four technologies makes the transition to an economy based on renewable energy even more challenging. The advances made in renewable energy power production to date have made little headway into the transportation sector. Although market penetration of electric vehicles into the markets for cars and light-duty trucks is growing exponentially in the United States, the major economic sectors of international shipping, freight railroads, and aviation are dependent on technologies that cannot be easily powered by wind or solar electricity generation. Although a push could be made for a fundamental technology change in transportation to displace the internal combustion engine or gas turbine, recent policy directives seem to prefer finding a renewable-based liquid fuel that can “drop in” to the existing technology mix.

A pathway ahead

If history can be any guide, then, it can be expected that the transition to a zero-carbon economy based on renewable energy sources using new conversion technologies will probably be slow, at best. Therefore, it would be unwise to expect new “breakthrough” technologies to suddenly solve the nation’s energy problems as society continues to depend heavily on incremental improvements to old technologies. Drop-in fuels might move faster because of compatibility with existing infrastructure that is typically slow to change, but those drop-in fuels have the downside of propagating older technologies.

Although the conversion devices listed in Table 1 have served the nation well for over a century, the most prevalent four systems involve a relatively inefficient step of converting thermal energy into motion. The relative inefficiency of that conversion yields significant waste heat and higher-than-necessary fossil fuel consumption. Thus, finding a way to bypass that step with some fundamentally different conversion approaches would be valuable. Today’s solar PV and wind turbines bypass the thermal conversion step, thereby avoiding the wasted heat, which is one reason why they are appealing. There are more potential pathways than can be comprehensively identified here, but a few are worth mentioning. The existence of bioelectric organisms, such as electric eels, that convert chemical energy into electrical energy directly presents an opportunity to harness this novel pathway through bioelectric power systems, such as microbial fuel cells, biological solar voltaic, piezoelectric protein skins, or photoactive fuel cells. Some people have even touted the controversial Casimir effect, which is claimed to convert atomic forces at submicrometer scale into forces that can induce motion. If proven, it could potentially be used as a pathway for using nuclear fuels to generate motion while bypassing the steam turbine.

In light of this analysis, we have several recommendations for energy policymakers in the United States and elsewhere:

• The EIA should track technology deployment as well as fuels. The EIA has remarkable capabilities for collecting data and performing analysis on energy consumption and flows. We recommend that the agency include in its Annual Energy Review a table and graph in the overview chapter similar to Table 1 and Figure 2 in this manuscript. Such information is already available in a disaggregated format throughout the rest of the annual report and could readily be assembled using reasonable assumptions (and inputs from a handful of other EIA reports) by EIA analysts. Although the existing body of information reported by the agency is already world-class, we believe these additions would yield additional insights about technology innovation in the energy sector.

• The United States should invest in R&D for new fundamental energy conversion technologies (in addition to fuels and incremental improvements in existing energy conversion technologies) and concentrate on energy conversion pathways that reduce waste heat. A vast preponderance of energy research has emphasized developing alternative fuels with the same old technologies, such as new biofuels for old spark-ignition engines, rather than seeking to reduce consumption through efficiency improvements or the development of alternative technologies. However, presumably there are many innovations in energy conversion waiting to be discovered, given the right support. These might be as simple as efficiency improvements to the existing technologies, innovations in the small-scale but newer conversion devices (including solar PV and solid-state devices such as fuel cells), the integration of existing systems (novel power cycles), or wholly new conversion pathways.

• Policies that can accelerate technology development and adoption at scale should be emphasized. Accelerating the deployment of renewable energy technology will be the key to making a nontrivial impact on global emissions and resource depletion before the end of this century. In particular, policies that focus on accelerating the transition and adoption of a more diverse set of conversion pathways should be encouraged.

Due to the nation’s continued dependence on 19th-century technology and the logical expectation that technology transitions will probably take many decades at best, it is imperative that policies to promote conversion technology development and commercialization be put in place soon. The transition is going to take a long time, so we might as well get started.

Changing the Way We Account for College Credit

For centuries, the United States has been the envy of the world in terms of its higher education system. But now we are largely coasting on a bygone reputation, obscuring the fact that high-quality, affordable college credentials are not getting into the hands of the students who need them most.

One of the greatest assets of America’s higher education system is that we try to provide broad access to college credentials. Instead of remaining content to have a handful of private institutions that largely served the elite, President Abraham Lincoln signed the Morrill Land Grant Act in the 1860s, providing support for the creation of our public land-grant universities. A century later, hundreds of community colleges were created to ensure open access for all who wanted to enroll in higher education after high school. As awareness grew about the prohibitive expense of a college education, we made sure that all students who wanted to attend could afford it by providing generous state subsidies and federal support such as the GI Bill and Pell Grants. These investments allowed unprecedented numbers of Americans to enjoy the benefits of higher education and helped make us the most college-educated country in the world.

But the tides are changing for our great system. We are slipping, fast. Once first in the world among the countries of the Organisation for Economic Co-operation and Development in terms of young adults with college degrees, the United States now ranks 14th. Whereas other nations’ young people are far more likely to have college degrees than their parents, the United States is on the verge of having an older generation that is better educated than the younger. This couldn’t come at a worse time. Technological development and structural shifts in the global economy mean that nearly two-thirds of U.S. jobs will require some form of postsecondary education in the next five years. One-third will require a bachelor’s degree or higher, and about 30% will require some college or an associate degree.

It’s not just the young who need higher education. Those who have seen their blue-collar jobs and membership in the middle class disappear are also yearning to learn the post-secondary skills essential for them to succeed economically. As routine work becomes increasingly automated, employers need workers with the skills necessary to handle complex and changing tasks and situations. A college credential is currently the easiest, if not necessarily the most accurate, proxy for those skills.

Yet even as college is becoming more essential, it is also becoming much more expensive. Tuition and fee increases have outpaced even health care costs, rising by more than four times the rate of inflation during the past 25 years. Students and families are covering those increases with student loan debt. Two-thirds of today’s students graduate with student loans, owing an average of $26,600. The nation’s collective student loan debt is more than $1 trillion, exceeding even our collective credit card debt.

Part of the problem is that our concepts of what colleges and college students look like have not kept pace with the realities. The collegiate archetype—a well-prepared 18-year-old ready to move into a dorm and study full time at the same college for four years, all of it paid for by mom and dad—is now the exception, not the rule. And as for the bucolic residential campus experience? That, too, is an exception. About 80% of today’s students are commuters. Nearly 45% of undergraduates attend community colleges. Nearly 60% attend two or more institutions before they graduate. More and more students are taking some or all of their courses online. In sum, students today are more likely to be older, working, attending part time, and learning outside of traditional credit-bearing classrooms than students in the past. Their lives demand a much different and much better kind of education.

Because many of today’s students are juggling work and family, higher education needs to be more responsive to these scheduling and financial constraints. It also needs to recognize the different experiences and goals that these students bring to the table. Right now, most colleges treat all students the same, as empty vessels that can only be filled within the confines of the college classroom. Compare two hypothetical students pursuing a bachelor’s degree in criminal justice. The first is an 18-year-old with no work experience whose interest in the criminal justice system comes from watching Law and Order reruns. The second is a 31-year-old high-school graduate who has hit a degree ceiling at the law firm where she has worked for 12 years. Colleges treat them the same; that is to say, largely like an 18-year-old.

The antiquated images of college and college students also rest on a false distinction between education and training, further harming students pursing education outside the framework of a residential college. More than 40% of students at community colleges are enrolled in noncredit courses, many of which are workforce-training courses requested by and developed in conjunction with employers. Many of these courses include highly sophisticated and complex subject matter, which is then assessed and certified by industry-normed exams. Although employers and students benefit from the training received from these courses, students do not receive college credit and thus miss out on the permanent, portable units on which the widely recognized and remunerated college certificates and degrees are built. This limits students’ ability to enter career paths and adjust to cyclical and structural changes in the economy over time.

Unfortunately, there are few incentives—the biggest one being federal financial aid—to address the needs of these students. Federal financial aid pays largely for time served rather than learning achieved in for-credit courses at established, accredited institutions. Granting credit based on seat time instead of learning gives credit where it shouldn’t and fails to recognize learning that happens outside of classroom walls. This hampers students’ acquisition of valuable degrees and credentials. It also creates structural disincentives for developing new means of giving students the kind of flexible, affordable, and effective education and credentials they need. Institutions are rewarded largely for inputs rather than outcomes. What are the grade-point averages or SAT scores of incoming students? How much have faculty published? Rarely is the question of what students are learning asked, let alone answered.

Why is this the case and what can we do about it? To answer, we need only invoke the old adage “you get what you pay for.” Right now, we are paying for time, not learning. In order to change that, we have to address underlying problems with the basic currency of higher education: the credit hour. This way of measuring student learning is putting our nation’s workforce and future prosperity at risk. That’s because when the credit hour was developed at the turn of the 20th century, it was never meant to measure student learning.

The curious birth of the credit hour

American secondary schools expanded dramatically around the turn of the 20th century, swelling the ranks of high-school graduates. But the extreme variation in high-school practice left many college admissions officers unsure as to what a high-school diploma meant. The National Education Association endorsed a “standard unit” of time students spent on a subject as an easy-to-compare measure. But the idea didn’t stick until later, when Andrew Carnegie set out to fix a problem that had nothing to do with high school: the lack of pensions for college professors.

As a trustee of Cornell University, Carnegie was troubled by poor faculty compensation. Professors made too little to prepare for retirement, leaving many to work far longer than was productive for them or their students. Carnegie decided to create a free pension system for professors, administered by the nonprofit Carnegie Foundation for the Advancement of Teaching. Not surprisingly, colleges were eager to participate. The foundation leveraged this excitement to promote another one of its goals—high-school reform—by requiring participating pension colleges to use the “standard unit” for college admission purposes. Colleges had nothing to lose and free pensions to gain, so the time-based standard unit, which became known as the Carnegie Unit, became the de facto standard for determining high-school graduation and college admissions requirements.

Carnegie’s pension system also spurred higher education to convert its own course offerings into time-based units to determine faculty workload thresholds to qualify for the pension program. Using the Carnegie Unit as a model, faculty members who taught 12 credit units, with each unit equal to one hour of faculty-student contact time per week over a 15-week semester, would qualify for full-time pension benefits.

Soon, the credit hour would become the fundamental building block of college courses and degree programs. The move to time-based units however, was never intended to be a measure of student learning. The Carnegie Foundation made this quite clear when discussing its “unit” in its 1906 annual report, in which it was explicitly stated that in the counting, the fundamental criterion was the amount of time spent on a subject, not the results attained.

But colleges did not heed this caveat, and it’s easy to understand why. The standardized nature of credit hours makes them convenient for a number of critical administrative functions, including determining state and federal funding, setting faculty workloads, scheduling, recording course-work, and determining whether students are attending college full time. The problem is that over the years, the credit hour has also come to serve as a proxy for measures of learning. Most importantly, college degrees came to represent the accumulation of credit hours, typically 120 to earn a bachelor’s degree.

More time does not equal more learning

College degrees are still largely awarded based on time served rather than learning achieved, despite recent research suggesting that shocking numbers of college students graduate having learned very little. The 2011 National Research Council study Academically Adrift found that 45% of students completing the first two years of college and 36% completing four years of college showed no statistically significant improvement over time on a test of critical thinking, complex reasoning, and communication skills. A U.S. government study found that the majority of college graduates could not do basic tasks such as summarize opposing newspaper editorials or compare credit-card offers with varying interest rates.

Perhaps time is still part of the equation; students should be spending two hours outside of class for every hour in class. But the reality is quite different. In 1961, two-thirds of students spent at least 20 hours a week studying outside of class. By 2003, the percentage had dropped to 20. But, theoretically, colleges supplement the credit-hour count of how much time students have spent in and outside of class with an objective measure of how much they have learned: grades. But it is hard to reconcile today’s grades with the research suggesting that poor learning outcomes are widespread. Whereas 15% of undergraduate course grades were A’s in 1961, today almost half are A’s. Nearly two-thirds of provosts and chief academic officers think grade inflation is a serious problem. Either college graduates have become much smarter over time—a possibility contradicted by all available research—or the function of grades in meaningfully differentiating and rewarding student learning has eroded.

Given these sobering findings, it is not surprising that employers are not particularly impressed with recent college graduates. Only one-third of employers say that college graduates are prepared to succeed in entry-level positions at their companies, and only about one-quarter said that colleges and universities are doing a good job in preparing students effectively for the challenges of today’s global economy. There is a curious disconnect between the widely held belief that American universities are great and the growing recognition that its graduates are not.

When an hour isn’t an hour

Perhaps the strongest evidence of the credit hour’s inadequacy in measuring learning can be found in the policies and choices of colleges themselves. If credit hours truly reflected a standardized unit of learning, they would be fully transferable across institutions. After all, an hour in Massachusetts is still an hour in Mississippi. But colleges routinely reject credits earned at other colleges, underscoring their belief that credit hours are not a reliable measure of how much students have learned.

Many students, however, believe that the credit hour is a standardized currency and assume that their credits will transfer from one school to the next. This is an unfortunate and costly assumption. Take the case of Louisiana community college students. Until recently, students with an associate degree typically lost between 21 and 24 credits when transferring to a four-year state school. That’s a year of time and money lost. Given that nearly 60% of students in the United States attend two or more colleges, the nontransfer of credits has huge individual, state, and national implications.

The government must be much more active in encouraging competency-based education and highlighting the competency models that use or could use the credit hour to receive financial aid.

Yet millions of credits are awarded that lead to degrees where very little, if any, learning is demonstrated. It is no wonder that the federal government has recently begun to weigh in on the credit hour. Because the cornerstone of federal financial aid, the credit hour, doesn’t represent learning in a consistently meaningful way, it is hard for the government to ensure that taxpayer dollars are well spent and that students are getting degrees of value. The problem has been exacerbated by two intertwined trends: the steady increase of online education and the growth of the for-profit higher education industry.

The concept of seat time becomes more difficult to measure when students aren’t in actual seats. From 2002 to 2010, the percentage of students taking at least one online class rose from under 10% to 32%. Online education fueled for-profit education most dramatically; students enrolled in for-profit programs increased more than 300% between 2000 and 2010. Federal policy fueled this boom, by removing a requirement that at least one-half of an institution’s students be enrolled in face-to-face courses to be eligible for financial aid.

One of the primary appeals of online classes is the flexibility they provide to students juggling work and family responsibilities. Online classes are often asynchronous, meaning that students don’t all gather in front of their monitors at the same time each week. Nor do students need to spend the same amount of time in front of their monitors; in many cases, students work at their own pace, going quickly through concepts they have mastered and lingering on those they have not. Although a boon to working students, online courses fit awkwardly with the seat-time basis of the classic credit hour and have become increasingly problematic for education regulators, particularly for the federal government.

Many online and for-profit colleges, as well as colleges of all kinds, are heavily financed by federal student financial aid dollars. In 2012, the U.S. Department of Education gave out more than $187 billion in grants, loans, and other forms of student financial aid, an increase of more than $100 billion in annual aid in just the past 10 years. And the building block of this aid is the credit hour.

Until recently, credit-hour determination was left entirely up to colleges and their peer accreditors. Institutions assigned a number of credit hours to a course and accreditors reviewed the process of determining credits. If the accreditors signed off, the department would open its financial aid coffers to students at the institution.

This began to change in 2009, when the department’s inspector general found inadequate accreditor oversight of the credit-hour assignment processes. Although colleges typically offer three credits for a 15-week course, the inspector general highlighted an institution that granted nine credits for a 10-week course. When the institution’s accreditor flagged what seemed to be an excessive amount of credits, the institution simply broke up the course into two five-week, 4.5-credit courses. The accreditor called the policy egregious, but approved the institution anyway.

In response to the inspector general’s report, and to a growing concern over poor quality controls for federal financial aid, the Department of Education decided there was a need for a consistent, standard definition of a credit hour. This would be tricky: Define the credit hour too tightly, and risk reifying seat time and stifling innovation; define it too loosely, and students and taxpayers could be taken for a ride.

After much discussion and controversy, the department released its final regulation in 2010. The definition used the time-based measure devised nearly a century earlier to determine eligibility for Carnegie’s pension plan. But it also provided alternatives based on evidence of student work or student learning. Unfortunately, the alternative parts of the definition were largely overlooked, in part because the credit-hour definition was just one piece of a series of regulations designed to reduce fraud and abuse in the financial aid program. Thus, many in the industry still believe that their safest and easiest bet is to do what they have always done: use time, rather than learning, to determine credits.

If we accept that college-level learning can occur outside of traditional institutions, then why shouldn’t we accept that college-level credit could be granted outside of traditional institutions?

The 15-week, one-hour-in-class-and-two-hours-out definition of a college course is not just easy to measure; it is a long-established practice and convention. The credit hour may be an illusion—studies suggest that typical students work nothing close to two hours out of class for every hour in—but it is an illusion that everyone understands and agrees to believe. This is in stark contrast to agreements about learning outcomes. Although colleges and their accreditors claim that learning outcomes are an integral part of an institution’s DNA, the research findings on poor learning outcomes and rampant grade inflation, combined with the difficulty of credit transfer, tell a different story.

Learning from others

Fortunately, there are institutions that have long used learning, rather than seat time, to award credits and degrees, in addition to more recent efforts to try and measure learning. In the late 1960s and early 1970s, the Carnegie Foundation emerged again as a central player in new approaches to higher education. As adults supported by the GI Bill and more women entered or returned to school, it became clear that existing time- and place-dependent colleges were often ill-suited to serving these learners. Carnegie produced a series of reports emphasizing that adults were not simply older 18-year-olds; they had skills, knowledge, and educational needs that traditional students did not. Institutions needed a different approach: one that started with recognizing, measuring, and awarding credit for the high-level knowledge and skills adults had acquired through life and work experience.

Several new programs and institutions were created in the early 1970s to address the needs of adult learners. Ewald Nyquist, New York State’s commissioner of education and president of the University of the State of New York, proposed a Regent’s degree program to help those unable to attend traditional college courses with the opportunity to earn a degree. The program used exams and validation of credits earned at other institutions to help students more quickly and inexpensively earn their degree. Learning outcomes and degree requirements were made clear, and students could demonstrate what they already knew and then spend their time learning what they did not. In 1972, the program’s first associate degrees were awarded.

The program soon became a college and eventually became Excelsior College, whose motto is “What you know is more important than where or how you learned it.” Over the years, Excelsior has broadened the ways in which students can earn credits and degrees, adding demonstration of prior learning through a portfolio of projects and in-person or online classes. In early 2012, it announced a modern and inexpensive twist to its competency-based programs. For less than $10,000, students can earn a bachelor’s degree by using free online courses and materials and demonstrating their mastery of the subjects on exams designed by experts from across the country. Excelsior has the largest nursing program in the country, and its graduates do as well as those from traditional time-based programs on national licensure exams.

Exams are commonplace in many industries: Lawyers need to pass the bar before being allowed to practice, doctors must pass the boards, and many professions have specific licensure exams that certify that individuals have the minimum competencies necessary to be credentialed in a given field (for example, Cisco information technology certifications). In higher education, however, learning standards and assessments are largely devolved to the level of the individual course. Individual professors often set their own singular standards, deliver instruction, and then measure the students against these standards. And although grades may be high, evidence suggests learning is scant. This is not to suggest that higher education can and should be measured by one big test. But it needs to do a far better job of identifying and objectively measuring what students are expected to and can actually do. Exam-based credits already exist, in a limited scope, in higher education; they just happen to be geared toward high-achieving, traditional students. In 2012, 3.7 million high-school students took Advanced Placement (AP) exams in the hope of testing out of courses whose material they had already mastered. Although students may have difficulty transferring credits from one college to another, they are likely to face less resistance in receiving credit for AP courses, because institutions know what AP tests mean and are more likely to trust them. These credits are trusted because they are based on learning, not time.

The financial aid hurdle

Despite these innovations, Excelsior and the handful of similar institutions that started in the 1970s remain relatively unknown commodities. This is in large part because students who enroll in competency-based programs typically have not had a key benefit available to students at most other accredited institutions and programs: access to federal financial aid. Although students in Excelsior’s online classes are eligible, students in its competency-based exam programs are not. According to the federal government, these programs are considered independent study programs, because they lack regular faculty-student interaction. This concept has been at the heart of many federal aid policies, largely to protect students and taxpayers from unscrupulous diploma-mill operators. If we can’t measure the time that distance education students spend in class, according to this the thinking, at least we can measure the time they interact with faculty.

In the 1990s, a new institution, Western Governors University (WGU), found a way to overcome the financial aid hurdle. WGU was started by the Western Governors Association, a nonpartisan group of governors from 19 states who were grappling with how to prepare their residents to meet the states’ workforce needs. Populations spread over sparsely populated stretches of the West and rapidly growing urban areas in states such as Nevada and Arizona needed much greater access to higher education. Creating hundreds of new brick and mortar institutions was not financially feasible, nor was expecting that working adults would leave their jobs and families to attend institution hundreds of miles away. The answer was a fully online institution.

But access alone was not enough. The governors heard employer complaints about the quality of college graduates and wanted to be sure students learned what employers needed. The key was competency. Groups of faculty, scholars, and industry experts would define the competencies that students would need to demonstrate for each degree program. Graders unconnected to the students would then evaluate the competencies. This approach not only provided a consistent benchmark for the quality of the degree, it also allowed students to move through the material at their own pace. New students are assessed for competencies they already have, and a learning plan is created to help them master the rest. Students pay a flat rate of less than $3,000 for six months, during which time they can move through as many competencies as they are able. The average graduate earns a bachelor’s degree in 30 months and pays a total of about $14,000. Employers are pleased with WGU graduates: According to a survey conducted by Harris Interactive, 98% rate the graduates as equal to or better than those of other universities, and 42% rate them as better.

Although WGU was originally built to serve students in the western states, today it serves students across the country. States are contracting to create state-branded versions of WGU, and enrollment is growing by 35% a year. This growth is made possible largely by the fact that students at WGU are eligible for federal financial aid.

It may be surprising to learn that WGU uses credit hours to receive federal aid. It wasn’t supposed to. Given the problems that seat time would pose to WGU’s competency-based model, Congress created an alternative way for institutions to receive financial aid, one that used the “direct assessment” of student learning. But WGU never used this authority, choosing instead to work creatively within the confines of the credit hour. Students are required to master 120 competencies, not coincidentally the standard number of credit hours required for a bachelor’s degree. WGU has regular faculty-student interaction, but its faculty members don’t teach. They function as mentors, helping students access the instructional resources they need to learn on their own. WGU’s creative use of the credit-hour requirement and faculty-student interaction has helped it access federal dollars critical to its growth.

Promoting new education models

Although Excelsior and WGU have broken away from seat time, the vast majority of colleges have not. Indeed, current government policies, and the misperceptions of these policies, have made it difficult for them to do so. If the United States is to reclaim its position as the most-educated nation in the world, then federal policy needs to shift from paying for time to paying for learning.

Many of the tools needed to make this shift are available to federal policymakers right now. Surprisingly, the first tool the government should use to help move away from time is its own recent definition of the credit hour. Many institutions and accreditors either don’t recognize the flexibilities in the new definition, or they don’t know how to use them. The government must be much more active in encouraging competency-based education and highlighting the competency models that use or could use the credit hour to receive financial aid.

The second tool is the direct assessment provision created for WGU. WGU hasn’t used it, nor has anyone else. That may soon change. In October 2012, Southern New Hampshire University, which is creating a $5,000 entirely competency-based associate degree, became the first institution to apply to use this provision. Freed from the history and practice of the credit hour, direct assessment could help institutions think more creatively about measuring student learning.

Although government can and should help push the boundaries of what is possible, it will not change the fact that measuring time is easy and measuring learning is not. This poses a real danger to innovation, because there are too many unknowns to safely broaden access to financial aid.

Fortunately, the third tool is the authority to test policies with small, voluntary groups of institutions. The government could ask a series of questions and try to answer them in controlled experiments. Should financial aid pay to assess learning that happened outside of the classroom? For learning achieved on the job? For learning achieved in Massively Open Online Courses (MOOCS)? How much should be paid to assess this learning? What should be the proof of student learning? Successful experiments could yield answers that would pave the way for wide-scale adoption.

These three tools offer a tremendous opportunity to move away from seat time. But a high bar must be set lest we recreate the grade inflation and weak academic standards in the existing time-based system or open up the floodgates to billions of dollars in federal aid to unscrupulous operators. Demonstrated learning outcomes are the key to this endeavor. The government should provide guidelines that are broad enough to support innovation yet stringent enough to prevent abuse. At a minimum, these guidelines should include transparent, externally validated learning outcomes.

But although these three policy tools could be extremely valuable in accelerating the completion of meaningful, learning-based degrees, they have limits. No matter what eventually might be covered by these tools, they apply only to accredited institutions. This means that noninstitutional providers of learning, no matter how good the outcomes, will remain ineligible. A biotech company could create a high-quality, work-based training program whose “graduates” have learned more than most students with an associate degree in science, but unless this training is attached to an accredited institution, the learning outcomes won’t count towards a credential.

If we accept that college-level learning can occur outside of traditional institutions, then why shouldn’t we accept that college-level credit could be granted outside of traditional institutions? For now, the law is very clear on who can grant credit and who can receive federal financial aid: institutions of higher education only. Perhaps after a few rounds of experimentation with the credit hour, direct assessment, and experimental sites, policymakers will see value in awarding credit for learning, irrespective of how long it took, where it happened, or who provided it.

As higher education becomes increasingly necessary and expensive, measuring time, rather than learning, is a luxury that students, taxpayers, and the nation can no longer afford. Moving to transparent, competency-based education would shed light on what students are learning in the classroom. It would also help millions of students capture and validate learning from outside the classroom, meeting students where they are and taking them where they need to go. Students and taxpayers can no longer afford to pay for a time-based measurement designed to help professors qualify for pensions. If we pay for what students learn and can do, rather than how or where they spent their time, it would go a long way toward providing students and the nation with desperately needed, more affordable, and better-quality degrees.

From the Hill – Winter 2013

House, Senate Committees investigate meningitis outbreak

After a meningitis outbreak this fall tied to a Massachusetts compounding pharmacy, House and Senate committees held hearings to investigate what happened, and members on both committees clashed with Food and Drug Administration (FDA) Commissioner Margaret Hamburg over whether the FDA could have done more to prevent the outbreak.

More than 14,000 patients in 23 states were exposed to a rare form of fungal meningitis through contaminated steroid injections. As of December 12, 2012, 590 cases had been identified in 19 states, resulting in 37 deaths.

In September, clinicians at the Tennessee Department of Health (TDH) confirmed several cases of unexplained fungal meningitis and notified the Centers for Disease Control and Prevention (CDC). A subsequent investigation by the TDH, CDC, and FDA determined that the New England Compounding Center (NECC) in Massachusetts was responsible for the contaminated injections. A subsequent investigation by the FDA and the Massachusetts Board of Pharmacy (MBP) found numerous regulatory infractions, inadequate sterilization procedures, and vials of steroid injections containing visible black particulate matter (later confirmed as microbial contamination). The CDC then activated its Emergency Operations Center to contact potentially exposed patients, monitor diagnosis and treatments, and provide up-to-date information on the outbreak. NECC recalled all products currently in circulation and permanently surrendered pharmaceutical licenses for the facility and its three primary pharmacists.

At the hearings held by the House Committee on Energy and Commerce and the Senate Committee on Health, Education, Labor, members commended the TDH and the CDC for their responses to the outbreak. Beth Bell, director of the CDC’s National Center for Emerging and Zoonotic Diseases, discussed the importance of federally funded programs that trained clinicians on the ground and provided laboratory and surveillance capacity in preparation for such outbreaks. But members questioned whether the MBP and FDA provided adequate regulatory oversight of NECC.

As discussed at the hearings, in the 14 years that NECC had been in operation, the FDA conducted three series of inspections (each related to a separate complaint or incident), issued two reports of objectionable conditions, and issued a formal warning letter for numerous violations of pharmaceutical regulations. The MBP investigated a dozen additional complaints, issued four advisory letters and reprimands, and entered into a consent agreement with NECC to resolve numerous problems related to the misfiling of prescriptions and problems with drug potency and sterility that had previously led to patient infections. In July 2002, two patients in New York were hospitalized with meningitis-like symptoms after receiving steroid injections from NECC contaminated with bacteria. Subsequent inspections by the FDA again identified problems with drug potency and sterility at NECC, leading the FDA to recommend that the MBP prohibit NECC from drug development until these problems were resolved. The FDA warned that there was “potential for serious public health consequences if NECC’s compounding practices, in particular those relating to sterile products, are not improved.” But NECC continued its operations.

Much of the discussion at the House and Senate hearings focused on two factors underlying the outbreak: the definition of a compounding pharmacy versus a drug manufacturer and the regulatory oversight of the FDA and MBP.

“From the Hill” is adapted from the newsletter Science and Technology in Congress, published by the Office of Government Relations of the American Association for the Advancement of Science (www.aaas.org) in Washington, DC.

In her testimony, Hamburg defined traditional pharmacy compounding as the combining or altering of ingredients by a licensed pharmacist producing a medication tailored to a patient’s special medical needs. This can include reformulating a drug to accommodate patient allergies or altering dosage for children or the elderly. All parties involved in the hearings, including members of Congress, cited the benefits of this practice. Although the FDA maintains heavy oversight over drug manufacturing, Hamburg cited a lack of policy guidelines on pharmacy compounding. She said that Section 503A of the Food and Drug Administration Modernization Act of 1997 exempts compounding drugs from several provisions of FDA regulation, which have been challenged in court with conflicting results and leave the FDA with ambiguous oversight. In such situations, the FDA defers to state-level regulators, whose oversight is more clearly defined.

Hamburg proposed a tiered system of regulation that separates traditional compounding, which requires minimal oversight, from nontraditional compounding, and clearly defines state and federal regulation in each category.

Several members of the committees challenged Hamburg’s perspective, arguing that repeated inspections, reports, and especially the 2006 warning letter indicate that the FDA believed it could exercise regulatory oversight over NECC, and they questioned why the FDA had not continued to do so. Hamburg responded that FDA authority was “limited, unclear, and contested.”

In her testimony, MBP Commissioner Lauren Smith stated that the MBP had “primary responsibility for oversight” and that “troubling questions remain” about MBP investigations of NECC.

The House Committee has requested input from all 50 state boards of pharmacy on the balance between federal and state regulation of compounding pharmacies and has drafted a proposal for legislation to improve drug distribution security.

U.S. Weather Commission proposed

The University Corporation for Atmospheric Research (UCAR) and the Weather Coalition sponsored a congressional briefing, “Toward a U.S. Weather Commission: Protecting Lives, Livelihoods, and the American Economy,” during which a panel of experts called on Congress to create a U.S. Weather Commission to protect the American people from economic and personal harm.

The panel included John Armstrong, chair of the Committee on the Assessment of the National Weather Service (NWS) Modernization Program; William B. Gail, cofounder and chief technology officer of the Global Weather Corporation; Pamela G. Emch, senior staff engineer and scientist for Northrop Grumman Aerospace Systems; and Thomas Bogdan, president of UCAR.

The briefing was held in response to the National Academies report Weather Services for the Nation: Becoming Second to None. The committee that authored the report pointed out that the NWS has not changed its operations or structure since its last modernization in the 1990s. The report contains three core recommendations for the service: prioritize core capabilities, evaluate the organization’s structure and function, and improve collaboration with all parts of the weather enterprise, including the private sector.

According to the expert panel, normal weather events cost the United States $485 billion per year, and major storms in 2011 alone resulted in another $52 billion in damages, 8,000 injuries, and more than 1,000 fatalities. A U.S. Weather Commission would provide guidance to policymakers as they implement the Academies committee’s recommendations to ensure that scientists have the resources they need to improve forecasting capabilities and warning systems.

UCAR and the Weather Coalition suggested that Congress model the new commission after the U.S. Commission on Ocean Policy, a largely successful advisory panel that provided guidance on ocean policy issues to President Bush from 2001–2004.

Federal science and technology in brief

  • The Departments of Commerce and Labor have announced a new initiative, the Make It in America Challenge, which will provide $40 million in competitive grant funding to accelerate the trend of in-sourcing, whereby companies bring jobs back to the United States and make additional investments here. The competition will be opened through a Federal Funding Opportunity to be announced by the start of 2013. It will focus on projects that accelerate job creation by encouraging insourcing of productive activity by U.S. firms, fostering increased foreign direct investment, giving U.S. companies incentives to keep or expand their businesses in the United States, and training local workers to meet the needs of those businesses.
  • On November 27, the president signed the Whistleblower Protection Enhancement Act, which protects government employees from retaliation when disclosing evidence of gross mismanagement, gross waste of funds, or abuse of authority within the government. Of interest to the research community, the legislation includes language that protects against censorship related to research, including efforts “to distort, misrepresent, or suppress research, analysis, or technical information.”
  • An effort to pass cybersecurity legislation failed, after Republican leaders objected because Senate Majority Leader Harry Reid (D-NV) would not allow an open amendment process. The Senate also failed to pass cybersecurity legislation in August. At that time, Republicans were concerned that mandatory security standards in the bill would put unnecessary burdens on the private sector.
  • On November 12, the Congressional Budget Office (CBO) released an analysis of the cost to implement the Great Ape Protection and Cost Savings Act (S. 810). The legislation would prohibit invasive research on great apes, require permanent retirement for animals that are currently being used in research, and expand an existing sanctuary to house them in retirement. According to the CBO, it would cost the federal government $56 million over a four-year period.
  • Francis Collins, director of the NIH, announced on November 16 that he has decided not to move forward with a recommendation from NIH’s Scientific Management Review Board that the agency establish a new institute focused on substance use, abuse, and addiction-related research. Thus, the National Institute on Drug Abuse and the National Institute on Alcohol Abuse and Alcoholism will retain their institutional identities. Collins formed an internal task force to recommend how a merger of the two institutes could best be accomplished, but in the recent statement he said he concluded that “it is more appropriate for NIH to pursue functional integration, rather than major structural reorganization, to advance substance use, abuse, and addiction-related research.”
  • Reps. Randy Hultgren (R-IL) and Chaka Fattah (D-PA) introduced H.R. 815, which declares 2013 the “Year of the Federal Lab” and highlights the accomplishments of national energy laboratories and more than 100 other labs that function as federal research centers.
  • On October 5, the Animal and Plant Health Inspection Service (APHIS) and the CDC published the final regulatory rule to govern the use of select agents and toxins. The rule implements a tiered system for regulating select agents based on risk for misuse and danger to public health; for example, 11 agents have been designated as Tier 1 agents and thus will require additional security measures. The agencies removed a number of other agents and viruses from the select agent list. In addition, the rule mandates stricter inventory requirements and audit procedures to enhance security. The rule went into effect on December 4, 2012.

Staying in School: A Proposal for Raising High-School Graduation Rates

High-school dropouts fare substantially worse than their peers on a wide variety of long-term outcomes. On average, a dropout earns less money, is more likely to be in jail, is less healthy, is less likely to be married, and is unhappier than a high-school graduate. Yet dropout rates in the United States have remained mostly unchanged, at roughly 30%, during the past three decades. This problem disproportionately affects low-income and minority students. Nearly half of these individuals do not graduate with their class.

A growing body of research, however, suggests ways to improve high-school graduation rates and close the achievement gap. A key element is for all states to increase their minimum school-leaving age to 18. Many studies have found that this intervention significantly improves several long-term outcomes. More effort is also needed to keep students engaged in school, even at an early age. If states invest in effective support programs, they can further increase graduation rates and reduce future costs of enforcing compulsory-schooling policies. All of these interventions should be implemented with the goal of strengthening the nation’s primary education system to promote college attendance and improve career outcomes among youth.

Lifetime of challenges

High-school dropouts face daunting challenges. Skills and educational attainment are increasingly important in today’s economy, and individuals with the least education are faring particularly badly. Among recent dropouts, 16% are unemployed and 32% live below the poverty line. Dropouts with jobs earn an average of only $12.75 per hour, with the most common jobs found in the construction, food services, and landscaping industries. Labor-market outcomes remain bleak throughout life. Dropouts aged 50 earn an average of $16.50 an hour and are most commonly employed in construction, food services, and truck transportation.

The act of dropping out must be understood not as a single event but an outcome that begins with school disengagement, often long before the dropout finally decides to stop coming to class.

Dropouts face worse social outcomes as well. For example, 33% of recent female dropouts have given birth as a teenager, 13% of male and female dropouts are separated or divorced, 32% report being unhealthy, and 22% report being unhappy, according to data from the 2005–2010 waves of the General Social Survey, which is considered a reliable indicator of societal trends.

Several studies also link a region’s proportion of dropouts to its overall prosperity. Individuals earn higher wages if they work in regions with fewer dropouts, irrespective of their own level of educational attainment. Crime rates are lower, and civic participation is higher. For these reasons, the high-school dropout rate is sometimes used as a quality measure of schools and an appraisal of the skill level of the future national workforce.

With so much hardship associated with leaving high school before graduating, why do so many students decide to do it? Of course, there is no single explanation: Conflicts at home, urgent financial difficulties, and unexpected pregnancies are only a few examples. Some dropouts say they are too poorly prepared to complete. A majority of these individuals say they are unmotivated or uninspired to go to class. Dropouts are truant more often, experience more academic troubles, and record more failing grades throughout all levels of schooling than do their peers who graduate. Dropouts are more likely to be from households where parents are less active in promoting and helping with school. By the time students decide to leave, they often feel there is disconnect or lack of support between themselves, their parents, and their teachers. The act of dropping out, therefore, must be understood not as a single event but an outcome that begins with school disengagement, often long before the dropout finally decides to stop coming to class.

Many studies have found that youth are particularly predisposed to impulsive behavior, especially in situations involving immediate costs relative to long-term benefits. Similar forces seem to be at play for many students in their decisions to drop out of school. In hindsight, adults who dropped out almost universally express regret. In one study, 74% admitted that they would have stayed in school if they could make the same decision again. So although the reasons students disengage from school are important to understand and address, the basic fact remains that students miss out on long-term payoffs from doing so.

For decades, laws compelling school attendance have been implemented with the goals of raising educational attainment, reducing the number of dropouts, and addressing the problems myopic youth and disinterested parents have in choosing whether the student stays in school. The compulsory-schooling age sets the minimum length of time that students must spend in school before they have the legal option to leave. States generally set the laws covering compulsory attendance. The laws have been around for many decades—in some cases, for more than a century—and they have been updated periodically, sometimes increasing and sometimes decreasing the time, depending on the particular needs and desires of each state. The youngest age at which students are now allowed to leave school is 16 (although often with some exceptions), which is the case in 21 states.

Much scientific evidence supports the view that increasing the compulsory schooling age is socially desirable. As President Barack Obama said in his 2012 State of the Union address: “When students don’t walk away from their education, more of them walk the stage to get their diploma. When students are not allowed to drop out, they do better.”

National blueprint

Requiring states to establish compulsory-schooling laws set at age 18 is not, however, a silver bullet for addressing the problem of dropouts. But it could form the cornerstone of a suite of policies to reengage the most at-risk young students; establish the right expectations for students, their families, and educators; and provide a focus for related policies to improve educational outcomes. In this light, we propose a four-part national approach to address the challenge.

First, the federal government should educate states on the benefits of high-school graduation and encourage legislative action to increase the minimum age at which students are legally allowed to drop out of high school to 18 years.

Compulsory schooling and education in general are usually legislated at the state level. The federal government, as it has recently done, can encourage states to consider more-restrictive laws and grade states based on the extent to which they follow federal recommendations. Even if the federal government could impose a national minimum school-leaving age, it is clear, based on experience, that such legislation is not likely to be effective if buy-in does not exist at the regional levels. The federal government has a larger role in disseminating best practices and motivating policies from a cost/benefit perspective.

On balance, the evidence suggests that students with more compulsory schooling would do better across a wide range of lifetime outcomes. The federal government can work to ensure that these benefits are clearly communicated to state and local policymakers, with the understanding that the gains from promoting high-school completion through compulsory schooling outweigh the costs of implementing and enforcing such laws.

Some of the best evidence suggesting that high-school students gain, on average, from staying in school comes from historical changes in compulsory-schooling laws. The first empirical studies in this area dealt with increases in the minimum school-leaving age that occurred in the first half of the 20th century. These studies consistently found large gains in adult socioeconomic outcomes. In the United States, studies found that annual earnings are nearly 10% higher for students compelled to stay an additional year in school, and comparable results have been observed in the United Kingdom and Canada.

In addition, studies using various states’ recent changes to the minimum school-leaving age have found that each year of additional schooling that students receive lowers the probability by 3.6 percentage points that they will end up unemployed, lowers the likelihood by 5.5 percentage points of their being on welfare, and lowers by 8.1 percentage points the likelihood of their living below the poverty line. Among those working more than 25 hours per week, a year of compulsory schooling is also associated with a 10.7% increase in annual earnings. These results may be understated, because education earnings gaps tend to increase with age, and the studies focused on younger cohorts.

Individuals nudged to finish high school through mandatory-schooling laws also fare better than dropouts on outcomes other than employment and income. Compulsory schooling has been shown in some studies to lower overall crime and incarceration rates, although there is some evidence that increasing the minimum school-leaving age to 18 results in higher in-school violence. Compulsory schooling makes individuals healthier: High-school dropouts are more likely to use cigarettes and illicit drugs than are high-school graduates, and better-educated individuals tend to have slightly longer life expectancies. Compulsory schooling reduces the incidence of teen pregnancy and may even have positive effects on memory and other cognitive abilities. There are also documented broader consequences of compulsory education that make democracies more effective by increasing political interest and involvement. In addition, it can decrease intergenerational inequality in educational attainment: Parents with more compulsory schooling are less likely to have children repeat a grade or drop out of school themselves.

Studies have also demonstrated the effects of increasing the school-leaving age above 16 on education attainment. In a study of a sample of 20- to 29-year-olds, for example, for each year the dropout age was extended above 16, school attainment increased by an average of 0.12 years per student. High-school completion rates increased 1.3 percentage points, on average, from increasing the school-leaving age from 16 to 17, and 2.4 percentage points from increasing it to 18. Raising the school-leaving age also led to an increase in college enrollment rates by 1.5 percentage points, suggesting that many of those encouraged to stay on and complete high school take advantage of new opportunities by pursuing college. Using these estimates, increasing the school-leaving age to 18 for every state would lead to approximately 55,000 more students completing high school and 34,000 more students entering college per year.

Lessons drawn from these and other findings can form a compelling argument that may encourage states to examine their compulsory-schooling laws and make sure they reflect the best current scientific evidence.

Second, states should be encouraged to develop new programs to reengage at-risk youth.

Compulsory-schooling laws help establish social norms and expectations for minimum school attainment. But compulsion should be a last resort alongside other policies to promote engagement and foster an environment in which struggling students are encouraged and assisted to complete high school. States should be challenged to come up with innovative plans, relevant to their communities, to keep young students engaged and learning before they approach high-school ages where they actually drop out.

The decision to drop out of school often results from a much longer process of disengagement that begins in elementary school. Patterns of high absenteeism and lower performance by future dropouts tend to start as early as the third grade. Thus, policies that combat early disengagement may prevent at-risk students from falling into a downward spiral, in which missing school causes them to fall behind in their studies, which in turn makes them feel even less motivated to attend classes and puts them further behind. At young ages, truancy is more often related to parental issues. Addressing parent situations that keep children away from school while working with parents to improve conditions for children to cope with the social and academic challenges of school are ways to foster school engagement. Children tend to do better when parents set high expectations for them. Setting rules and helping with homework are ways that parents can encourage their children to adapt to school early and do well in the long term.

Parents also need to be actively involved through all levels of schooling. Although many parents become more involved on learning that their child is considering leaving school, they are often not aware of their child’s poor performance until it is too late. When school administrators and educators communicate more regularly with parents regarding their children’s performance, they provide a means for parents to take a more active role.

The school environment itself is obviously another strong determinant of whether at-risk students succeed. Students who are supported, motivated, and encouraged by their teachers, who regard their teachers as caring, and who receive guidance from their teachers usually like school. In contrast, dropouts often report leaving school because they did not get along with their teachers or classmates. Smaller class sizes or counseling and guidance programs for struggling students are ways to improve how students perceive their teacher support networks.

Recent evidence points to the importance of setting high academic expectations. Students should be made to feel that they are expected to complete high school, and that teachers and parents are there to help make that happen. Compulsory-schooling policies, in a broader context, exist to set minimum expectations about school attendance and attainment. A number of other interventions show promise in fostering expectations and engagement, even at an early age. Mentoring programs, especially, provide opportunities for administrators to directly interact with students and families, and they are relatively cost-effective. For example, Check & Connect is a program that sends support workers to meet with students and parents in urban middle schools to discuss attendance and academic performance. A number of randomized controlled trials have found that the program, developed at the University of Minnesota, leads to lower tardiness and absenteeism and increased graduation, as well as to increases in literacy and school completion. Of course, this intervention is not without cost—the price tag of Check & Connect is approximately $1,100 per student per year—but the long-term benefits of the program probably far outweigh its costs.

Third, state and local governments should improve the enforcement of new and existing laws.

Although a strictly enforced minimum school-leaving age should, in theory, cause every student to either remain in school until the requisite age or face a penalty, compulsory-schooling laws tend not to be strictly enforced, often for reasons of cost. For example, Los Angeles had 260 attendance counselors to cover 660 schools in 2002. Yet Boston had only seven truant officers in 2003, and Denver had only one officer in 2004. In the 1990s, Chicago eliminated all truant officers for budgetary reasons in favor of using mentoring programs to tutor and reengage moderately truant students. More resources clearly should be devoted to hiring more truant officers and attendance counselors.

Although the importance of involving parents in dealing with truant children cannot be overstated, punitive measures may be useful to curb absenteeism in cases where counseling is not as successful. Parents may be unwilling or unable to discipline children to attend school, especially for older adolescents who are more independent. In some cases, the parents may also suffer from problems of mental illness or drug addiction, which further complicates the matter. The credible (and possibly implicit) threat of fines and court hearings for parents may motivate them to make sure their children are attending school regularly. Community service requirements, fines, and misdemeanors sanctioned on students are other ways of targeting truancy in cases where absenteeism is a result of student delinquency. Imposing restrictions on driving privileges has also shown to be a successful deterrent to truancy. However, because fines and court hearings are resource-intensive, they should be used as last resorts after counseling or other corrective measures have been exhausted.

Also, exemptions to compulsory-schooling laws are common in many states. These exemptions permit adolescents to leave school when obviously appropriate. In many cases, dropping out is a result of teenage pregnancy, the need to care for a family member, or an immediate need to make money. States often address these concerns by requiring parental consent, coupled with a younger minimum dropout age, to ensure that basic educational requirements are still met. Having clear exemptions in place would give administrators flexibility to accommodate cases in which the costs of continuing are obviously too high.

These different approaches to enforcement will probably have different results. Researchers are only now learning about which approaches work best. For instance, a study is under way to explicitly test different approaches to enforcing compulsory-schooling laws in a sample of truant youth in Chicago, in the absence of any other enforcement of those laws.

Fourth, compulsory-schooling laws should be designed to promote college attendance and improve the career outcomes of students.

The motivation for a renewed focus on compulsory-schooling laws is the increasingly poor labor-market outcomes of high-school dropouts. In the past several decades, an increasingly competitive global economy and technological advancements have reduced the job opportunities available to less-educated, less-skilled workers and increased them for higher-skilled workers. This increased demand for skilled labor is reflected in the rise in earnings premiums of high-school and college graduates as compared with those of high-school dropouts. For example, the U.S. Census Bureau reported in 2005 that individuals who graduated from high school earn 1.5 times more than dropouts, and college graduates earn 2.7 times more than dropouts. Yet even high-school graduates have seen large declines in their average earnings levels and employment rates relative to college graduates during the past several decades. This suggests additional opportunities for improved social and economic success through college. Increasing high-school attainment should be regarded as part of a more general goal to make youth more competitive in the labor market.

Compulsion should be a last resort alongside other policies to promote engagement and foster an environment in which struggling students are encouraged and assisted to complete high school.

Although traditional education is clearly highly prized in today’s labor market, the goal of compulsory-schooling laws is not to try to shoehorn all students into that model. Its purpose should be to prepare the nation’s youth for a range of opportunities, including college and careers. To that end, nontraditional programs that have a clear track record of proven results should be eligible to fulfill the requirements of compulsory-schooling laws. For instance, vocational training and school-to-career programs may also be effective ways of providing disengaged students with alternative school options for developing practical, hands-on learning and useful skills for success in the labor market.

Career academies—programs in high schools focused on preparation for postsecondary education and the transition to work—enable students to combine standard classroom learning with on-the-job work experience. These programs are primarily located in urban schools and serve a cross section of students, including many at-risk minority students. They offer small learning communities that combine academic and technical curricula around a career theme to enrich learning. Monthly earnings are higher for career academy graduates, particularly for males, than for similar students who did not participate in the programs. More important, these gains most consistently accrued to students classified as being at the highest risk of dropping out of school. The option to receive assistance in finding work after school also entices many students to graduate on time, because the financial benefits of remaining in school are made more salient. Per-student costs of school-to-work programs can be similar to those of mentoring programs and may even be repaid by graduates through subsequent payroll taxes.

Furthermore, compulsory schooling may be one way of encouraging youth to pursue higher education, even though such laws do not explicitly mandate it. There is evidence that some individuals nudged to complete high school become more interested in college or view higher education as less daunting an obstacle than when they were younger. Some studies have also found that mandating that students complete the college application and financial aid process may lead to increased college enrollment.

In addition to communicating the importance of continued education and mitigating absenteeism, mentoring programs and parent-teacher outreach initiatives also provide opportunities to help students choose courses and plan their learning paths to achieve their long-term career goals. It is important that high schools offer an array of course options to keep students interested in learning and prepare them for college. Vocational training, such as student-work programs, may be useful for providing students who would otherwise typically not go on to college with real-world work experience, which may in turn open doors to a successful career.

Weighing costs

Economic evidence points to sizable financial and nonfinancial benefits, on average, for students from increasing the minimum school-leaving age. But although compulsion and truancy prevention affect particular groups of students more than others, costs and resource burdens affect schools, administrations, and states on a wider scale. This raises the question of whether such costs are justified or offset by the observed gains of at-risk groups.

On the benefits side, for example, is the estimated 10% increase in annual income, on average, from encouraging a student to stay a year longer in school. This means that the lifetime earnings increase from finishing high school and joining the labor force at age 18 rather than exiting high school at age 16 is approximately $226,700. After correcting for the fact that much of this income comes long after the high-school years, this sum is equivalent to a one-time payment of $94,300 at the age of 16 (when individuals are facing the decision of whether to drop out). There are also the many benefits of compulsory schooling that are difficult to quantify, such as reduced teenage pregnancy, improved individuals’ health, reduced dependency on public support programs, decreased crime rates, and increased voting and political involvement. Thus, beyond the increase in earnings of would-be dropouts, there are broader benefits to the students and to their families and communities.

On the cost side, more truant officers or caseworkers would probably be needed, although it is unclear precisely how many. According to estimates made by the Maryland General Assembly in 2012, if each worker makes $85,000 per year (the average in the state) and one worker can monitor 40 students, that is an increase of more than $4,000 to monitor each additional student for two more years.

Also, accommodating tens of thousands of students across the country who otherwise would have dropped out will entail direct costs from some combination of hiring more teachers, building new schools, or increasing class sizes. Per-pupil spending in the United States is roughly $12,300 per year, based on 2011 data from the National Center for Education Statistics. If accommodating each new student costs this amount, a state would pay almost $25,000 to keep a 16-year-old dropout in school through graduation. In reality, however, adding students to the education system is likely to increase costs by far less than the average rate per student. Some new schools and classrooms may have to be built to accommodate the would-be high-school dropouts who remain in school, but most of the infrastructure already exists to support these students. Actual costs of education, then, may be closer to $10,000 or $15,000 for each additional student.

Beyond these direct costs, there may be indirect costs, many of which are difficult to quantify. For example, increasing the number of students in public schools could lead to larger class sizes if schools accommodate these students without hiring new teachers or building new classrooms. Furthermore, some of the students who would remain in school as a result of the law change may be disruptive to their peers because they are among the least-enthusiastic students. This is most likely to affect students who are already struggling, who are most at risk of being distracted and falling into bad habits. There are also concerns that the incidence of crime and violence in schools might rise because of the increased attendance of unhappy and unwilling teenagers. Finally, many schools are already pressed for funds, and the increased financial burdens from compulsory schooling would divert resources from other valuable uses.

Positive bottom line

The best estimates suggest that the economic benefits that accrue to students who graduate more than offset potential costs. Although there is not an off-the-shelf compulsory-schooling program, a comparison can be made by hypothesizing a program that has substantially more teachers and classrooms to hold class size constant, as well as substantially more truant officers to deal with potentially unenthusiastic students. An upper estimate for these costs is roughly $28,800 for each student who is compelled to stay in school for an additional two years (although actual costs may be significantly lower). Thus, per student, the combined quantifiable benefits of increasing the compulsory-schooling age appear to exceed the costs substantially.

It also should not be overlooked that the vast majority of the costs are incurred in the actual education of would-be dropouts, as opposed to the enforcement of any laws. But any intervention that succeeded in reducing the dropout rate by a commensurate level would entail those same direct education costs. Compulsory schooling, then, is only expensive insofar as it is successful in keeping students in school, which, the economic evidence suggests, is a worthy goal.

Furthermore, programs that target disengagement among at-risk students at an early age will not only increase high-school graduation rates and the ensuing benefits that compulsory schooling brings; they also have the potential to significantly lower many of the associated costs. For instance, states will have to devote fewer resources to enforcing compulsory-schooling laws if relatively less-expensive programs such as Check & Connect curb disengagement early in students’ academic careers. These programs also may decrease the number of disruptive students in the public school system, which would minimize some of the potentially negative peer effects of compulsory schooling.

Plus, better-educated students will help drive today’s economy, which is searching for such employees, and the higher income-tax revenues that will result will partially offset states’ costs and provide a worthwhile return on investment. All told, states may find that many of these support programs are sound investments for increasing their population’s education level and economic outcomes in the most cost-effective way.

The Hidden Risks of Energy Innovation

Recent years have been disappointing for U.S. advocates of aggressive action on climate change. Efforts to pass comprehensive cap-and-trade legislation, which would have promoted the deployment of clean energy by making dirty energy more expensive and thus cut U.S. emissions, failed spectacularly. Global climate change negotiations, expected by many to deliver a legally binding deal that would sharply constrain global greenhouse gas emissions, have done nothing of the sort. Meanwhile, warnings of dangerous risks in the climate system have mounted.

People who understand that the climate problem is serious have reacted by grasping for new ways that government can lead on this front. Out of this searching, a big new idea has emerged: The world can cut through the dead-end politics of climate policy by focusing on clean technology. At home, rather than penalizing dirty energy, government would step in to help make clean energy cheap, a much more positive agenda that its proponents argue would eventually encourage mass adoption of low-carbon energy sources. Internationally, virtuous competition to win a clean energy race would replace the tired and unproductive squabbling that has marked efforts to agree on emissions constraints. Cheaper clean technologies would also mean that developing countries, where most future energy demand will come from, will fuel their economies cleanly, since cheap low-carbon options will become the ones that also enhance economic growth.

Alas, the turn from regulation to innovation is not a magic recipe for eliminating conflict over domestic or international policy, or even for significantly reducing it. Instead, it will create new fights in new spheres. This is not a reason to reject a big technology push as part of a serious climate strategy; climate change needs to be confronted, and conflict is almost certainly endemic to serious climate policy. Nonetheless, before policymakers place their bets on technology policy, they would do well to better understand the opportunities for conflict that lurk there. If they do, they will realize the limits of technology policy and will more likely pursue a modest but constructive approach. If they do not, the more likely outcome is a drive that tries to do too much with technology policy. But just like the maximalist efforts to solve every climate problem with cap-and-trade and an international treaty, that overstretch is likely to beget failure.

Promise and problems

The political logic motivating calls for a new focus on technology in U.S. domestic policy is straightforward. Americans do not like regulation, as is evident in their reactions to cap-and-trade, but they are enthusiastic about innovation and technology. They are averse to constraints, and whereas cap-and-trade was squarely about limits, technology promotion is about expanding options and possibly stimulating economic growth. Moreover, many see greenhouse gas regulation as being about creating losers, most notably in fossil fuel industries and those who depend on those industries for energy; technology promotion, in the new narrative, is about creating new winning industries and large constituencies who can gain from the options they provide.

The international case for a focus on technology is similarly straightforward. Traditional climate negotiations are about spreading the pain of emissions reductions. Each country fights to ensure that it is spared onerous obligations and that others bear as much burden as possible. Focusing on technology would sidestep that fight. If clean technology becomes cheaper than dirty fuel, then all countries will want to adopt it, and there will be no burden to be shared. Climate diplomacy is also immensely complicated. Negotiations over rules for measuring emissions and schemes for trading greenhouse gas credits take diplomats deep into difficult details. Technology, by contrast, appears to harness markets straightforwardly to help spread low-emissions behavior without the need for major international coordination or technical negotiation.

Much of this appears persuasive, yet there is good reason to believe that it is wrong. Americans may not like regulation, but they do not appear to like government spending or taxes either. Yet without new taxes, spending, or regulation, government has no significant tools with which to promote clean energy innovation. Technology promotion can superficially appear to be purely about encouraging growth—cheaper energy should mean more economic activity—but that ignores the cost of promoting clean technology in the first place. Indeed, ill-conceived government efforts to cut the cost of clean energy would simply spend taxpayer funds without producing any real world payoff. In addition, technology policy, if it is to succeed in dealing with climate change, will inevitably create losers alongside its winners. Cutting emissions requires not only using more wind, solar, and nuclear power, it also requires using less coal and oil.

Further problems loom on the international front. Technology promotion, just like emissions cuts, requires dividing a pie. The only difference is that the pie is a new one. Only so many wind turbines, solar panels, and nuclear reactors can be sold into the international market, and countries (and their firms and workers) will fight to maximize the fraction that are theirs. Moreover, a race to develop clean energy will not necessarily spawn a race to deploy it. Technology policy can make clean energy cheaper, but not necessarily cheaper than fossil fuel alternatives, particularly existing coal power plants whose capital costs are already sunk. Policy interventions such as cap-and-trade or regulatory mandates, with all the political challenges they entail, will still be required to tip the scales.

None of this means that clean technology policy should be neglected. After all, it is not as if other dimensions of climate policy are problem free. Policymakers will need, however, to confront the challenges of crafting effective technology policy head on. They will also need to take special care to maximize the odds that their policies are well designed. The fact that people are skeptical of government efforts to promote economic and technological change increases the importance of ensuring that technology promotion efforts are seen to succeed.

Navigating the home front

Domestic policy design faces one central question: Where should government intervene? This question is easiest to settle at two extremes. Technology in the earliest stages of research will be hard pressed to find significant commercial support. Ideas are leaky even with robust protection of intellectual property; investments by one firm in creating knowledge will often redound to the benefit of others instead. Economists have thus long identified spending on research as a place where governments can and should intervene constructively in markets. Those making U.S. energy policy should take that advice.

At the other extreme is mature and commercially established technology. Here, government should largely avoid subsidizing deployment. Some will argue that intervention is justifiable on environmental grounds. Subsidies to deploy wind turbines, for example, may make sense because of the value of wind energy in reducing emissions. But this is not a fiscally feasible path in the long run. Subsidizing the deployment of clean technology at a scale that would actually create deep cuts in U.S. greenhouse gas emissions would eventually cost hundreds of billions of dollars each year.

It is much more difficult to determine what makes policy sense in between. That territory, where developers build full-scale first-of-a-kind projects to demonstrate the feasibility of new technologies, scale up their businesses, and learn how to reduce costs through early stage commercial deployment, is known as the “valley of death” for good reason. There are only so many investors willing to risk significant amounts of money on unproven technologies that might take many years to turn a profit, particularly when the long-term policy picture is so unsettled and the ultimate market is unknown. Without government intervention, many promising inventions inevitably wither here on the vine.

This problem has been recognized in many areas of technology, but it is particularly acute when it comes to energy. During the past 60 years, and particularly over the last 30, the venture capital industry has become critical to sheparding technologies through the valley of death, removing much of the need for government intervention. It has been particularly successful in areas such as information technology and biotechnology. Encouraged by the recent explosion of Silicon Valley interest in clean technology, many people who think about energy assume that venture capitalists will play the same role in clean energy as they have in other fields. But most areas of energy are a poor fit for the venture capital model. Venture capitalists deploy small to moderate amounts of capital over periods of about three to five years on technologies whose intellectual property can be protected through legal means such as patents and copyrights. This business model is a perfect fit for things like software development, which does not require extraordinary amounts of money, where product lifetimes are short and thus new products can penetrate markets quickly, and where intellectual property is relatively straightforward to protect.

Most types of energy technology, though, do not fit this bill. First-of-a-kind biofuels plants cost hundreds of millions of dollars, and new-design nuclear reactors or coal-fired power plants that incorporate carbon capture and sequestration cost billions. Capital stock turnover is slow, with vehicles lasting a decade or more and power plants lasting on the order of half a century, which makes rapid market penetration impossible. And intellectual property law is often poorly suited to protecting important advances, including demonstration of the commercial viability of well understood but heretofore undeployed technologies, or innovative business models that bring down the cost of deploying fuel and energy-saving technologies. All of this means that venture capital cannot be counted on to take energy technologies through the valley of death, and that government will need to consider intervening much more seriously.

Yet the prospect of substantial market intervention immediately sends the climate problem back whence technology was supposed to liberate it: the realm of grubby politics and ideology. Just as there are fundamental philosophical divides over the role of regulation in the U.S. economy, which shaped the cap-and-trade debate, there are big ideological differences over the legitimate role of the U.S. government in intervening deeply in markets for new technology. Indeed, the turn to technology may make ideological fights worse. One can plausibly argue that cap-and-trade does not pick winners and losers in the economy; it simply (and subtly) shifts the playing field. In contrast, an active effort to help technologies bridge the valley of death will inevitably require supporting individual firms and technologies, precisely the sort of behavior that opponents of government intervention often find most troubling. Remember the uproar when the Solyndra solar energy company defaulted on its $535-million federal loan. Moreover, just as the cap-and-trade debate set off a flurry of industry efforts to shape the system to their interests, so too will a big technology push; the only change will be in the set of supplicants. Indeed, the problem may be worse. With cap-and-trade, fights over the distribution of the spoils were largely inconsequential to the policy’s environmental effectiveness. In contrast, successful rent seeking from technology companies may mean that large amounts of capital are steered to dead end firms and ideas, leaving less for real prospects. That would hurt the policy’s performance.

An active effort to help technologies bridge the valley of death will inevitably require supporting individual firms and technologies, precisely the sort of behavior that opponents of government intervention often find most troubling.

New sources of international conflict

The challenges entailed in crafting effective technology policy do not end at the water’s edge. Any U.S. technology policy will inevitably need to have two international goals: blunting climate change by maximizing the global deployment of clean technology and strengthening U.S. competitiveness by maximizing the global deployment of U.S. clean technology. Anyone who believes that the second goal can be discarded is not taking domestic politics seriously; anyone who would neglect the first is not talking about climate policy. Yet the two directions will often conflict. Moreover, even when they don’t, fights over which country will get a bigger piece of the clean energy market will be inevitable, particularly if countries succeed in making that market grow.

Economists have a straightforward response to these dilemmas. Technology spreads through international trade and investment and by the free and secure flow of ideas. Consumers in one country buy new products made in another; firms set up factories overseas that use their new technology; companies license innovations developed elsewhere. Economists argue that if this international market for technology is allowed to work freely, the United States will maximize its economic gains while also protecting the climate. If that means that wind turbines are made in India, or solar panels are made in China, rather than both being made in the United States, that is simply the market telling the United States that it can make more productive use of its talent and resources. At the same time, the United States can benefit from cheaper clean technology made overseas.

There is much to this argument, but it has important limits. For starters, the decision to trade freely isn’t a unilateral one. U.S. policymakers may decide that a liberal system of trade and investment will help square their climate and economic objectives, but China may choose not to play along. U.S. policymakers will then be forced to select among unattractive alternatives, whether acceding to large market distortions or choosing to launch a trade war. Either way, free market dogma will not tell them what to do.

Moreover, the same market failures that exist domestically—underinvestment in R&D, a lack of capital and patience in the valley of death—also exist internationally, suggesting that optimal economic outcomes will require governments to correct those flaws. Yet there is no reason to assume that such interventions will make the United States better off. Just as policies that improve domestic economic efficiency need not benefit all firms, it is perfectly plausible that international policies that look sensible from a global economic standpoint will not be beneficial to all countries. For example, the most efficient policy for moving new energy technology through the valley of death may encourage U.S. inventions to be manufactured in China, which may not ultimately benefit the U.S. economy. Moreover, even if a given policy will ultimately benefit Americans, voters may distrust it and demand measures that more explicitly assure them a big slice of the pie. Alas, the results are unlikely to be ideal.

There is also a deeper problem that is peculiar to clean energy. Efforts that succeed in reducing barriers to international trade and investment, and hence greasing the wheels of technology diffusion, may end up gutting markets for clean energy. Why? As long as clean energy is more expensive than dirtier alternatives, other domestic policies will still need to tip the balance toward investments in clean options. Such policies, though, are often advocated on the grounds that they are good for national competitiveness. U.S. policymakers, for example, have often made the case for climate policy by arguing that it will strengthen U.S. industry. In India, the government has advocated efforts to massively increase the use of solar energy as a way to boost domestic solar firms and has set aside a large chunk of the solar market for them in order to make that point. In China, efforts to deploy large numbers of wind turbines are justified not only as helping with energy security, but also as providing a foundation for a strategic export industry.

Yet a successful effort to knock down trade and investment barriers may mean that domestic firms that would have benefited from policies that encourage clean energy deployment lose out. With those firms no longer poised to gain from domestic policy that promotes clean energy, the political forces advocating low-carbon policies would suffer. Ultimately, demand may be gutted. That would be a loss both for U.S. exporters—a big slice of a nonexistent market is not worth much—and for efforts to combat climate change.

A modest way forward

A shift from cap-and-trade and regulation to technology and innovation will not banish politics or diplomacy from the climate scene, either at home or abroad. But there are steps that prudent policymakers can take to insulate a technology strategy from the greatest risks and to maximize the odds that it will deliver significant payoffs. The key to a successful technology strategy is modesty. A maximalist strategy will run into far more roadblocks and conflicts—and is more likely to fail—than one that seeks to make serious contributions without attempting or claiming to solve the entire problem by itself.

There is no better place to learn this lesson than from the ill-fated climate strategies of the past few years that proponents of a new technology push want to replace. At home, cap-and-trade morphed from a modest attempt to control greenhouse gases into a core fix for a moribund U.S. economy and its dependence on foreign oil. This made for an implausible case. At a very basic level, Americans found it difficult to believe that charging them more for electricity could revive the U.S. economy. It also required advocates to make sweeping claims that turned off many voters. Many Americans instinctively recoil from claims that regulatory policy is the route to economic growth; they are far more conditioned to the more modest claim that regulation is necessary for the environment. To be certain, the death of cap-and-trade was overdetermined. Economic uncertainty and poisonous partisanship were leading factors in its demise. But overwrought claims that it could solve the entire U.S. emissions problem alone, all while delivering a critical blow to economic stagnation, did not help its fate.

The international drive for an all-encompasing treaty suffered from similar problems. A global treaty with commitments to big emissions cuts and significant penalties for noncompliance should have been recognized early on as a nonstarter. States have little certainty as to whether they will be able to deeply reduce their emissions over the coming decades, and since they do not know whether they can actually make deep emissions cuts, they have always been unlikely to make costly promises to that end. Moreover, by choosing the maximalist goal of negotiating a global agreement, states added additional and unnecessary complication. Twenty countries are responsible for about 80% of global emissions, but strategists sought a deal among 192, leaving themselves vulnerable to the machinations of states from Venezuela to Sudan.

A wise technology strategy should be robust yet restrained in its ambitions. At home, it should start with a push on research, while avoiding massive subsidies to the widespread deployment of mature or nearly-mature technologies, which would stretch the government’s mission and open it up to reasonable accusations of fiscal irresponsibility. The recently created Adavanced Research Projects Agency–Energy (ARPA-E) is the sort of effort that makes sense in this vein; large economywide subsidies to the deployment of mature renewable energy or nuclear plants are the kind that do not.

Government funds should be used to help move technologies through the valley of death, since without them, the technologies needed to cost-effectively deal with the climate problem are unlikely to materialize. This argues in favor of some direct support for early deployment of technologies such as advanced biofuels and carbon capture and sequestration. But strategists should stop short of advocating federal government spending later in the technology development process, where government is more likely to merely be subsidizing private firms rather than constructively shifting the course of markets. As part of the economic stimulus, some of this was done in areas such as battery manufacturing, but this was done primarily in the context of economic policy, not energy policy. As policy turns to long-term transformation of the energy system rather than efforts to jumpstart a depressed economy, such interventions are harder to justify. Moreover, they have been tainted by public skepticism of the 2009 stimulus; anything that connects a new push on energy technology to that now unpopular policy will make a new energy strategy much harder to sell.

Many will ask how such a limited strategy can solve the U.S. emissions problem. That misses the point. It cannot and should not try to. Regulation and other incentives, including cap-and-trade, carbon taxes, or a clean energy standard, that do not cost the government massive sums of money will still need to play a critical role in promoting fundamental transformation in the energy system. These tools, particularly cap-and-trade, are politically toxic right now, but the climate problem is a multidecade challenge, and there is no reason to assume that the old toolkit is permanently useless. Indeed, a core goal of domestic technology policy should be to make regulation and other incentives easier. Anything that helps cut the cost of technology will reduce the economic burden of other policies when they are eventually pursued, making them more politically palatable. Seeing technology policy as part of a broader strategy, rather than something that must solve the entire climate problem alone, will allow policymakers to focus it on those areas where it can make the biggest contribution and avoid others that are most likely to get it into trouble.

A similar philosophy should shape U.S. technology strategy abroad. The United States is pursuing international cooperation on research with countries such as China and India, and should continue. In particular, its recent efforts to create joint clean energy research centers with both countries are laudable. It should complement that with a push for liberal markets for international trade and development in clean technologies (something that has been part of the Doha round and has been pursued through Asia-Pacific Economic Cooperation) and for more effective intellectual property protection. That task will be most difficult with China. The United States has already been forced to launch a controversial World Trade Organization suit against Chinese solar panel subsidies, but it should be attempted nonetheless. To maximize its odds of success, the United States will need to work with others since it is not the only country hurt by protectionist policies.

That said, as with the domestic front, the United States should draw limits. U.S. spending on big, near-commercial technology projects in other countries, such as full-scale clean coal plants in China, should be approached cautiously, as it is likely to raise political ire. It is also more likely to be a substantive bust, if only because the big price tags involved in individual projects will make it much tougher for the United States to spread its bets across multiple initiatives. U.S. diplomats should also stop short of opposing every barrier to trade and investment in clean energy that other countries throw up. Some of those barriers will be necessary to maintain demand for low-carbon technologies. For example, the United States can tolerate some local-content requirements in India’s solar initiative without suffering real economic harm. Moreover, by picking their fights, U.S. negotiators will be more successful in those that they choose to pursue.

Just as on the domestic front, technology strategy will need to be one piece of a broader international puzzle, lest policymakers try to do too much with it and inevitably fail. Yet unlike cap-and-trade and robust regulation at home, there is essentially no chance of revival for robust global treaty efforts complete with strict targets, timetables, and sanctions for noncompliance any time soon. Fortunately, there is also less need. Countries’ decisions about emissions have always been much more about domestic interests than foreign policy. A less formal approach that focuses on coordinating policies across borders, already under way as the result of some modest progress at the global climate talks, as well as innovative efforts to coordinate technology promotion such as the Clean Energy Ministerial (CEM) process, may be the best that is possible in the near future. More robust international efforts will eventually be required, but for now, maximizing room for domestic policy development, including through the right approach to technology, is the most important thing that can be done.

This approach to policy may leave those who worry most about climate change cold. When faced with a massive problem, people naturally grasp for an all-encompassing solution that promises salvation. Yet such schemes invariably reveal themselves to be mirages, and overwrought efforts to realize them too often backfire. Wiser policy will involve modest moves forward on multiple fronts, including technology. It would be tragic if policymakers chose a different course and replaced one overburdened climate strategy with another.

Not a Hundred Millionaires: The National Academy and the Expansion of Federal Science in the Gilded Age

Shortly before the nation’s centennial, the physicist Joseph Henry, long admired abroad as well as at home, a onetime Princeton professor and the secretary of the Smithsonian Institution for a quarter of a century, was stirred to reflect: The surprise is not “that science has made comparatively little advance among us, but that … it should have made so much.”

A good deal of what Henry noticed arose from the new dispensations for science by the federal government. The federal initiatives had begun in 1862, when the first Civil War Congress passed the Morrill Act, with its grant of lands to the states for colleges and universities that would offer studies in the agricultural and mechanical arts. In 1863, the same Congress created the National Academy of Sciences, a private agency with the public role of advising the government on policy-related technical issues.

No less important, in the years after the Civil War—the Gilded Age, as the era is known—the federal government authorized multiple new research efforts and agencies, enough to constitute a huge expansion in federally sponsored scientific research and scientific services to the American people, including four different topographical and geological surveys of the far West. By the mid-1880s, federal science had accumulated sufficient prominence and power to warrant that telltale hallmark of arrival in the capital: the scrutiny of a lengthy congressional investigation.

Henry’s surprise at the expansion was altogether understandable. Before the Civil War, astronomy, natural history, and natural philosophy (the term then used for physics and chemistry) were part of the required college curriculum. But science was in the main taught only at an elementary level along with the classics and theology as a tool of liberal education rather than for its own sake. In all but a few institutions—notably the two military service academies, Rensselaer Polytechnic Institute, and curricularly segregated parts of Harvard and Yale—its uses for practical purposes were by and large not only ignored but held in contempt.

Federal science was minuscule. It comprised only the U.S. Coast Survey, which had been established in Jefferson’s administration; the Army Corps of Engineers; the Naval Observatory; and an agricultural section in the Patent Office. All were devoted primarily to research for practical purposes. In contrast, the Smithsonian Institution, a hybrid of public and private enterprise, was committed to the advancement as well as the diffusion of knowledge, and Henry emphasized the progress of abstract knowledge as best he could within his severely limited budget. Notable among his achievements was to make the agency a center for the gathering of far-flung meteorological data by telegraph, the collation and analysis of the data, and the issuance of daily weather reports. Practical or abstract, government research was a pauper of national investment, and scientists were scarcely considered indispensable to the conduct of federal business.

Measured against its point of departure on the eve of the Civil War, the expansion of federal science in just the few decades after the firing on Fort Sumter was huge, and it was game-changing for the scientific enterprise in the United States.

Unlike the expansion that occurred after World War II, this 19th-century transformation had little to do with national defense. Its root cause was that the nation’s circumstances were changing in ways that established a mounting need for scientific and technological knowledge. Agriculture was growing ever more dependent on mechanization, chemistry, and biology. The United States had expanded to the Pacific at the end of the Mexican War in 1848, and the encouragement of settlement in the new territories entailed the mapping of the new lands and assessment of their natural resources, the establishment of military posts and transportation routes, the building of railroads, and the stringing of telegraph wires. An increasing number of Americans recognized the need for an infrastructure of technical education and research that would foster the progress of the emerging knowledge-based and technologically driven economy.

These Americans made up a loosely connected but determined coalition for training and knowledge. They included the educational reformers who found a champion in Senator Justin Morrill of Vermont, eager to invoke the resources of the federal government to democratize higher education and orient it toward meeting the needs of the emerging age. At the center of the coalition were the leaders of American science, a small, internationally respected group, among them, in addition to Joseph Henry, Harvard’s Louis Agassiz, a native of Switzerland and a brilliant student of rocks and fossils; and the geophysicist Alexander Dallas Bache, Benjamin Franklin’s great-grandson, the head of the Coast Survey and an authority on terrestrial magnetism. In concert and separately, these men had shaped the course of American science since the 1830s, when Henry had returned from a kudos-filled tour of Europe vowing to Bache that “the real working men … of science in this country should make common cause … to raise their own scientific character.”

By the end of the 1850s, Henry and his allies had been further energized in their ambitions by the rising intellectual ferment in science, especially the recent theories about the evolution of the physical Earth and, in 1859, Charles Darwin’s extension of evolutionary theorizing to Earth’s living inhabitants. Eager to enlarge the contributions that Americans might make to these great subjects, the advocates of science wanted to improve its condition in institutions of higher learning and enlarge its opportunities in the federal government.

During the decade before the Civil War, defenders of limited government, especially Southerners, feared any expansion of federal power. They had resisted most initiatives along these lines, just as they had blocked federal investment in the building of transcontinental railroads, the promotion of industrial development, and Justin Morrill’s land-grant bill. But after the South’s secession from the Union in 1861, the way was cleared for the East and the burgeoning middle West to form an imposing alliance determined to satisfy pent-up demand for federal promotion of Western settlement and national economic development under the leadership of the recently formed Republican Party. This was a Republican Party that was committed to industrial and economic expansion and was comfortable with the federal government’s assisting the achievement of that goal across a broad spectrum of activities, including investment in human capital and the capital of knowledge.

The advocates of science, their agenda fully in resonance with Republican aims, found considerable support in Washington during the boom years of the Civil War and beyond. The Morrill Act sailed through Congress with large majorities and so, in 1862, did the elevation of the agricultural section from its subordinate place in the Patent Office into an independent department. In short order, training in the agricultural and mechanical arts formally became a part of public higher education in the United States, breaking apart the standard uniform curriculum for liberal education by allowing for specialized study at advanced levels in science and engineering, among other subjects.

The Morrill Act goaded all American colleges and universities into becoming democratically accessible “engines of modernity” in the apt phrase of a discerning student of their development. Indeed, private institutions of higher education, embracing a similar curricular reform after the Civil War, gave comparable new attention to the sciences and their uses—so much so that Ralph Waldo Emerson remarked “that a cleavage is occurring in the hitherto granite of the past and a new era is nearly arrived.”

Among the prime movers behind the creation of the National Academy of Sciences were Agassiz and Bache. Agassiz the academic saw in a national academy an institution that would raise the quality of science in the United States by granting the imprimatur of membership not to men of mere learning but only to men of original scientific achievement. Bache, the longtime federal scientist, felt the need for an institution of authoritative scientists that would safeguard public policymaking in an increasingly technical age from charlatans and pretenders.

Joseph Henry opposed the creation of a highly selective national academy, suspecting that it might be considered “at variance with our democratic institutions” and might become “perverted … to the support of partisan politics.” However, without Henry’s knowledge, Agassiz, Bache, and their like-minded allies drafted legislation to establish the academy and enlisted support for it from Senator Henry Wilson of Massachusetts. A leader in the new Republican Party, Wilson was a passionate abolitionist, an ardent advocate of strengthening American prestige and progress, and chairman of the Senate Military Affairs Committee. In all, he was a skilled and respected politico, who later became vice president in Ulysses S. Grant’s second term. If Wilson spoke, the Congress listened, and when he introduced the bill for the academy on March 3, 1863, the last day of the session, it cleared the House and Senate by voice vote.

Bache was elected the first president. Agassiz, elected foreign secretary, was thrilled, holding that the nation’s men of science now had a “standard for scientific excellence.” Henry was far from pleased, but he agreed to membership in the academy, attending its first regular meeting in January 1864 and joining 21 other members at a White House reception with Lincoln, whom he greatly admired. Lincoln returned the compliment, noting that the Smithsonian Institution “must be a grand school if it produces such thinkers as he is…. I wish we had a few thousand more such men.” Henry eventually came around to the idea of the academy, accepting the presidency in 1867, upon Bache’s death. As president, Henry did not press for federal funds and kept the academy scrupulously out of politics. The membership ceiling was removed so that five scientists could be added to the rolls each year. Patronizing research by rewarding success with titles and pensions was the European way and unacceptable in America, Henry noted. Still, he concluded that an “intelligent democracy” could properly bestow honors for achievement, and the creation of the academy had opened in America another “avenue for the aspirations of a laudable ambition.”

An inauspicious start

Henry, however, was overly optimistic. During the post–Civil War years, a variety of circumstances militated against the National Academy’s playing a significant role in American science. Its financial resources were severely limited, insufficient to publish more than an occasional proceedings and obituaries of its members. It met in the Smithsonian, having no headquarters of its own, and the meetings were poorly attended. American scientists were dispersed over the eastern and midwestern United States, and Californians felt hopelessly isolated. Federal scientists tended to dominate the academy’s affairs, particularly elections to the ranks. Scientists in Cambridge and New Haven, wary of the centralization of science, worried that it was developing into a Washington clique.

Federal requests for the academy’s advice were few and far between. But in mid-1878, Congress appropriated funds for the organization to prepare a report on the consolidation of the multiple territorial surveys and provide an overall plan for assessing and mapping the nation’s territories.

Within months, the academy submitted its recommendations. In one part, it called for the unification of the four geological and topographical surveys, two of which were run by the military, into a single civilian geological survey. In another part, it incorporated the scientifically informed proposals for management of the public lands in the West that had been advanced by the famed geologist John Wesley Powell in his Report on the Lands of the Arid Regions of the United States. Not yet a member of the academy, Powell had written his report for his superiors in the Department of the Interior. Nevertheless, by embracing its proposals, the academy found itself embroiled in a political firestorm, what amounted to the first contest over climate science and public policy in the nation’s history.

Fame had come to Powell as a result of his intrepid journeys as the head of several successive western surveys into territories hitherto unknown to most Americans. Despite having lost his right forearm at Shiloh during the Civil War, he had scaled Long’s Peak, braved the canyons of the Colorado River, and returned with valuable knowledge of the region, including maps. Powell was a staunch supporter of material development and economic opportunity, but he was also a social reformer. He empathized with the Native American tribes in the West and initiated what became a major program of ethnology to learn about their languages and customs. He was also convinced that scientific knowledge had to be deployed to accommodate democratic social progress to the realities of the land and its limits. Powell knew the American West well. Partly because of the Smithsonian’s meteorological and weather program, to which Joseph Henry had given him access, he was aware that the West beyond the 100th meridian—a region accounting for some 40% of the land in the continental United States—experienced too little rainfall to sustain the conventional 160-acre homesteading agriculture that had long marked public land policy. Drawing on these data, Powell knew, in short, that most of the good public land, the kind “a poor man could turn into a farm,” had already been sold.

However, a number of earth scientists, including some who were also veterans of the western surveys, thought differently. Drawing on European theories and their own experience, they contended that settlement and cultivation, especially the planting of trees, would transform arid regions, notably the Great Plains beyond the 100th meridian, into fertile, loamy expanses bathed in rain, like the agriculturally fecund Mississippi Valley. During the 1870s, as settlement spread westward through the Great Plains, rainfall happened to increase, providing these arguments with a seeming plausibility and later giving rise to the phrase: “Rain follows the plow.”

Powell dismissed these claims in his Report on the arid lands, noting the “many conjectures and hypotheses” that had been advanced to account for the increased rainfall. He went on: “Many have attributed the change to the laying of railroad tracks and construction of telegraph lines; others to the cultivation of the soil, and not a few to the interposition of Divine Providence in behalf of the Latter Day Saints.” But he had to add that “in what manner rainfall could be affected through the cultivation of the land, building of railroads, telegraph lines, etc., has not been shown.”

In his Report, Powell called for an end to the apportionment of public land in 160-acre tracts. He urged instead the scientific classification of the land in ways that allowed for large sections suitable for dry-land ranching and irrigation farming. Their contours would be shaped by their access to water and their size—up to 2,500 acres—and would be determined by the use to which the land could be put: mining, grazing, farming, or irrigation.

The academy’s recommendations, both for the consolidation of the surveys and Powell’s reform of the public land system in the arid region, were made the basis of a congressional bill in February 1879 that Representative Abram S. Hewitt, a respected iron manufacturer and friend of science, commended to the House, declaiming that the plan came from “the highest scientific authority of the land.” The measure was hailed by Easterners on both sides of the aisle as wise, realistic, and scientifically compelling.

But representatives from the western states and territories, ambitious inhabitants of a developing and settler-hungry region, ripped into the proposal. Had scientists said that the land was too arid for small farms? Congressman Martin Maginnis from the Montana Territory cried that theorists had pronounced the homestead system dead before, yet settlers had gone west and, “practical men” all, had “seen the capabilities of this land which had escaped the notice of our scientists and statesmen.” Did the proposal enjoy the imprimatur of the highest scientific authority in the land? Congressman Thomas MacDonald Patterson from Denver, Colorado, argued that the National Academy of Sciences had “never published but one work, and that was a very thin volume of memoirs of its departed members.” He added to the laughter of the House, “And if they are to continue to engage in practical legislation, it would have been very well for the country if that volume had been much thicker.”

In the end, Congress combined the surveys into a new U.S. Geological Survey (USGS), but it left the 160-acre-based public land system intact. Congressman Dudley C. Haskell, a Yale graduate from Lawrence, Kansas, did not mind giving science its “little appropriation,” but he had to say: “Now, if you want a geographical survey, if you want a lot of astronomical figures, if you want a lot of scientific material, then organize your geographical surveys and authorize them to get out there and dig and hunt bugs and investigate fossils and discover the rotundity of the earth and take astronomical observations. But if you please, while you are there acting in the interest of science and in the interest of professional bug-hunting, leave the settlers upon our frontier alone.”

In fact, time would prove that Macginnis and his settlers had the matter wrong and Powell and the academy had it essentially right. On the Great Plains during the 1880s, rain ceased to follow the plow, ruining numerous farmers; and irrigation would prove indispensable to agriculture in the region. The battle would not be the last to join contested theories of climate with passionate convictions about political economy.

Yet for all their disagreements with Powell and the academy, the congressmen from the West were not antiscience. What they objected to was the deployment of scientific authority to make a radical change in their region’s land-distribution system. Their interests were economic, based on expectant hopes for rapid population growth and development. Although they had their climatology wrong, they valued federal science highly for all the research, analysis, and information it was providing for the practice of agriculture and the exploitation of natural resources.

Indeed, there were multiple indicators of Congress’ ongoing enthusiasm for an expansive federal science enterprise. In 1871, it had authorized the Coast Survey to encompass geodesy by measuring the curvature and gravitational force of Earth at its surface throughout the continental United States, which meant along an arc equal to about a full eighth of the circumference of the globe; in 1878, it affirmed this sizable expansion of the agency’s activities and renamed it the U.S. Coast and Geodetic Survey. A key contributor to this effort was Charles Sanders Peirce, the famed founder of philosophical pragmatism. During his 30 years with the survey, Peirce produced a stream of brilliant papers on formal logic and philosophy while earning an international reputation for his scientific work, especially the development of highly accurate methods for measuring gravity with a pendulum.

Congress also remained highly enthusiastic about an earlier initiative in the expansion of federal science, the U.S. Weather Service, which it had established in 1870 in the Army Signal Corps. The Weather Service, which took over the meteorological work and methods of the Smithsonian, served the needs of agriculture and the military, not to mention the American people in general. Army and civilian observers at stations around the country would wire reports of local conditions to the service’s Division of Telegrams and Reports, housed in the quarters of the Army Signal Corps on G Street in Washington. There, the next day’s predictions were drawn up, then telegraphed across the United States. The reports were published in newspapers and tacked up at local post offices. Many farmers learned of cold waves, sunshine, or storms from signal flags hoisted on a passing train or a public building (the flags would remain a commonplace of the American landscape until the days of radio). The service also supported research in meteorology, encouraging the enlistment of college graduates and hiring civilian scientists. To improve weather prediction, they explored the dynamics of storms and tornadoes and investigated the mysteries of the atmosphere.

Congress’s creation of the USGS, in 1879, was a clear vote of confidence for what the civilian surveys had done to date. Perhaps the best-known of them apart from Powell’s venture down the Colorado had been the Geological and Geographical Exploration of the Fortieth Parallel, which followed one of the expected routes of the transcontinental railroad and was headed by Clarence King. Then only 25 years old, King was handsome, well connected, and in Henry Adams’ judgment “the best and brightest man of his generation.” King, who had studied at Yale under some of the country’s leading geologists, staffed the exploration with first-rate scientists. Their work was fruitful scientifically, and much of it was also economically pertinent, especially the geological investigation of the region in the neighborhood of the rich Comstock silver lode.

King’s survey included artists and photographers. Their work, displayed in popular lectures with stereopticons and published in handsome books, provided untold numbers of Americans with their first visual impressions of the West. The illustrations of the canyon country, the mesas, and the high peaks, by turns majestic, cathedral-like, and forbidding, held them in thrall, much as pictures from space would rivet Americans a century later.

King was appointed the first head of the new USGS, but he resigned the post in 1881 and was succeeded by John Wesley Powell. His exploring days over, his body paunchy but his beard still a bristling red, Powell fired the entire operation with enthusiasm. He employed topographers, geologists, and paleontologists; he farmed out work to university consultants, a number of them at the land-grant colleges and universities in the Midwest and the West. Through its consultantships and field expeditions, the survey provided important research opportunities for many American geologists. Its access to the Government Printing Office enabled it to publish hundreds of bulletins valuable to both basic and economic geology. No disdainer of Washington politics, Powell could usually find a position in the survey for the relative of a well-placed congressman. He also distributed the survey’s attractively, sometimes lavishly, illustrated publications around the capital. With its numerous publications and consultants spread through 19 states and territories, the survey made its influence felt in numerous academic localities, and its work gained world-class distinction.

By the mid-1880s, the federal government, with its venerable Naval Observatory, expanded Coast and Geodetic Survey, new Weather Service, and jewel of the USGS had become the home of much of American science. The Government Printing Office, which since 1870 had issued hundreds of memoirs for the scientific bureaus, was the nation’s principal publisher of research. Relative to population, more scientists were working in the capital than in any other city, including Cambridge, Massachusetts. Washington scientists formed a congenial community, gathering at meetings of the various scientific societies then proliferating in the capital and conjoining at the Cosmos Club, which Powell, Henry Adams, and friends had founded in 1878 for people devoted to science, art, and literature. It was an intellectually stimulating community, and its members felt secure in the knowledge that through their work under federal auspices they were contributing to the advancement of science, the nation’s economic and cultural development, and the overall welfare of the American people.

They were thus jolted in the mid-1880s when the showpieces of the expanded federal scientific establishment, including the USGS, were subjected to the scrutiny of congressional investigation. No mere formality, this was an extended inquiry by a joint bipartisan commission of the House and Senate headed by the Republican warhorse Senator William B. Allison of Iowa, the powerful chair of the Committee on Appropriations. The investigation had been prompted in part by charges of scandal, favoritism, and wastefulness in the scientific agencies. Treasury auditors were reportedly investigating claims that John Wesley Powell himself had been doing useless research, overpaying favorites, publishing lavish books on irrelevant subjects, buying his way into the National Academy with patronage, and giving away federally owned fossils to a paleontologist at Yale.

But the fundamental issue was not scandal. It was whether the scientific agencies had developed beyond the limits of what the federal government should support. Critics contended, variously, that the scope of federal science had grown too broad and that the scientific agencies were engaging in abstract work of no utility—for example, meteorological studies that might not lead to more reliable weather prediction or paleontological investigations that seemed to satisfy no public want. Such objections came from states-rights southern Democrats, newly returned to Congress from their exile, and from some northerners, including several leading scientists. In the outspoken view of the critics, much of the research done by agencies such as Powell’s should be left to the states and to private enterprise.

Federal scientists vigorously defended their work before the Allison Commission, insisting, as one put it, that they were “not fomenting science”; they were doing practical work for practical purposes. The National Academy, once again asked for an opinion, urged that federal research should be removed from politics by consolidating all of it under an apolitical Department of Science.The commission rejected that recommendation, but Senator Allison and his Republican allies remained strong friends of federal science, and after two years, federal research emerged from the probe unscathed.

A few years later, Powell and federal science would suffer severe cutbacks at the hands of southern Democrats, western Republicans angered again by his insistence on managing public lands in accord with the need for irrigation, and even some eastern Republicans in thrall to cost-cutting in the face of hard economic times. But the setbacks to federal science were temporary. At the opening of the 20th century, with the depression of the 1890s ended and the United States emerging as a world power, the expansion of federal science resumed. Within a half century, it would reach levels undreamed of by Powell, and the federal government would call on the advice of the academy to a degree that Alexander Dallas Bache could hardly have imagined.

The foundations of that future had been laid in the Civil War, when Congress established the academy, and during the Gilded Age, when the Allison Commission sanctioned the purposes and practices of an expansive federal role in science as proper functions of government. It was only a short step in time and principle from the affirmations of the Allison Commission to Congress’s passage of the Hatch Act, in 1887, which created federally funded agricultural experiment stations at the land-grant colleges and universities and marked the beginning of an explicit commitment to federal patronage for scientific research in the academic world. Powell had explicated the issue of the federal role in research before the commission with clarity and passion. Not even “a hundred millionaires” could support the current research of the federal agencies, he declared, adding that the progress of American civilization should not have to wait on the philanthropic inspiration of a hundred rich men. Surely, he insisted the national government should support and publish whatever science might advance the welfare of the American people.


To read the footnoted version of this article, click here for the PDF.

Moneyball

The chorus of praise that has greeted Paula Stephan’s How Economics Shapes Science is well deserved. I am only echoing other reviewers by describing it as learned, insightful, eloquent, and timely. Stephan, an economist at Georgia State University, brings broad knowledge and rigorous analysis to the complex web of social and political forces that are inescapable—but too rarely discussed—part of the scientific enterprise. Stephan’s goal is not to propose any simple economics-based solutions to the conduct of science but to enrich our understanding of the social system in which science operates. This is not reductionist money-talks economics, but a sophisticated discussion that considers, for example, how the interplay of intellectual curiosity, professional reputation, institutional culture, and financial realities influences the researchers’ actions.

Yet, Stephan does pay attention to the bottom line and peppers the book with interesting information about science’s costs. For example, a typical mouse for research costs $17-$60, but a designer mouse with a particular disposition to diabetes or obesity could cost $3,500. A single researcher uses many mice and can spend more than $200,000 a year in mouse maintenance. The author tenders much more information about what it costs to maintain a faculty member and how that differs between private and public universities, across disciplines, and up and down the hierarchy from grad students and postdocs to full professors.

Stephan also cites more general data that has implications for the research enterprise’s overall effectiveness: 50% of highly cited physicists work in a country other than where they were born; 30% of papers by U.S. authors with more than one author have a non-U.S. coauthor; 6% of scientists and engineers produce 50% of published papers.

Throughout the book Stephan encourages readers to think more deeply about all aspects of the research environment. Prizes matter, but many choices determine how effective they are: Is the honor for a single discovery or a life’s work? Who decides on the priority goal to be incentivized? Do too many new prizes steal thunder from established prizes? We know that university scientists are spending more time consulting with the private sector, but we lack detailed data on how extensive this is, how much money is involved, how the nature of the consulting might be changing, and how this form of technology transfer stacks up against other routes. Collaboration is becoming more necessary in research, but we have yet to figure out related questions, such as how to assign intellectual property rights to discoveries made by teams and to acknowledge group achievement in an award system that has traditionally lionized individuals. Generalizations about research are always suspect because its nature differs by discipline and project, varying from the mathematician working alone in her office to the army of scientists and engineers collaborating at the Large Hadron Collider.

The physical capital of research is also changing, and the result is both greater democratization and increased stratification. Advances in information technology have made it easier for researchers to gain access to data from telescopes and gene sequencers. On the other hand, some research requires direct access to very costly lab facilities and equipment that might be available at only a few institutions and thus to only a handful of researchers. Should this changing landscape influence which types of research deserve government support?

In sifting through the details of how research is conducted today, Stephan does identify a few common problems. The procedures for managing the research enterprise were developed for a much smaller system and need to be updated to today’s scale. Slow growth in available funding has made the system too risk adverse, and the research world has become increasingly inhospitable for young researchers.

Focus on people

The most enlightening and disturbing chapters address the conditions and prospects for young people trying to launch careers in science. These chapters need to be read in full to appreciate the complexity of Stephan’s analysis, and she is too wise to propose a simple solution. But she raises serious questions about the economic efficiency of our current system for training future scientists and makes a strong case that the system is fundamentally organized to serve senior researchers at the expense of the young.

This imbalance prevails partly because 44% of U.S. PhD degrees in science and engineering are awarded to students on temporary visas and that almost 60% of postdocs are on temporary visas. Stephan devotes an entire chapter to the plight of many young people who come to the United States from other countries for graduate school and postdoctoral training. These talented young people contribute hugely to research during their training, and most find a way to stay in the United States when their training ends. But Stephan points out the risks of depending on foreign talent. Many countries are devoting resources to attracting scientists and engineers back home, and a sudden change—such as a Chinese ban on study in the United States—could be very disruptive. Stephan does not reach firm conclusions on whether foreign students are “crowding out” native citizens from science and engineering careers, but she does find evidence that the influx of foreign-born postdocs is depressing wages.

When Stephan returns at the end of the book to the relationship of science to economic growth, she aims to disabuse readers of simplistic notions of how university research leads, albeit slowly, to private sector innovation and increased economic productivity. She explains that innovation entails much more than new science and technology, that university intellectual property practices can sometimes slow technology transfer, that the training of researchers who will work in industry is of more immediate value than the research done in university labs, that spillover effects from research done in firms play a critical role. This is not to deny the important contribution of publicly funded research conducted in universities. But it helps us see that making the most of the national investment in research is not a simple matter.

Stephan’s review of some of the possible inefficiencies in the U.S. research system at least sets the course. Universities could be overinvesting in laboratory facilities and equipment to attract more federal research dollars, the nation could be training more PhD scientists and engineers than needed to perform the research we can afford to fund, and the ups and downs of federal research support play havoc with individual careers. The author reminds us that senior scientists and university administrators do respond to incentives, so that policies should ensure that incentives to individuals and institutions align with national goals.

To help young people considering careers in science make good choices, she would require universities to report the career outcomes for the people they train. Further, she would make research less dependent on the work of trainees because this would reduce the incentive for PIs to hire too many grad students and postdocs. She summarizes the argument made by many that training grants and fellowships are a preferable source of support for grad students and postdocs but acknowledges that empirical support for this case is lacking.

As researchers advance in their careers, Stephan would like them depend less on outside funding to maintain their university positions and to have more access to support for collaborative and interdisciplinary research. And she would like to end the “doubling” rhetoric popular among some science advocates. The rapid doubling of the National Institutes of Health budget, she avers, resulted in inefficient use of resources. A more reliable and predictable level of funding would avoid that problem and disrupt careers less.

Stephan ends by speculating about several questions without definitive answers. As she observes, we have no reliable way of calculating what percentage of gross domestic product should go to research, but the nation could probably benefit from spending more than it does now. The heavy U.S. emphasis on biomedical research raises obvious questions about balance, and she is inclined to shift some resources toward the physical sciences and engineering. On the questions of what size grants are most efficient or whether it would be better to fund individuals for long periods rather than projects for short periods, she is agnostic.

A subtitle to Stephan’s book could be: Consider the Counterfactual. Stephan warns against thinking that because something desirable happened—a new medical treatment was developed or a labor-saving device produced—and because it can be linked to some previous scientific research, that we did the right thing in supporting that research. The desirable event might have occurred without the earlier research funding or we might have stimulated even more and better events if we had spent that research money differently. Astutely, the author calls her final chapter “Can We Do Better?”

Archives – Winter 2013

The NAS Building at Dusk

National Academy of Sciences Building, 2101 Constitution Avenue, NW, Washington, D.C. © 2012 Maxwell MacKenzie.

The National Academy of Sciences is a private, non-profit society of distinguished scholars. Established by an Act of Congress that was signed by President Abraham Lincoln in 1863, the NAS is charged with providing independent advice to the nation on matters related to science and technology. The NAS celebrates its 150th anniversary in 2013.

The National Academy of Sciences Building, depicted here, opened in April 1924. It closed in 2010 for the restoration of its historic core, renovation of later additions, and installation of new communications networks, and electrical, heating and ventilation equipment. It reopened in April 2012.

Global Bioethics: Hopes, Fears, and New Voices

During the 1990s, James Grifo, a physician and researcher at New York University, had been working to develop a technique to help treat certain kinds of infertility. Although in vitro fertilization (IVF) treatments had been successful for many of his patients, IVF could not help women whose eggs were genetically sound and could be fertilized, but were not viable enough to grow into a healthy embryo. In such cases, Grifo imagined it might be possible to remove the nucleus from a donor egg from a healthy woman, replace it with the gene-carrying nucleus taken from the patient’s egg, and then implant the reconstructed egg into the patient’s uterus where it would continue to develop. Because the implanted egg would retain the mother’s DNA, she would give birth to a biologically related child.

Although the idea had never been tested, it gave hope to one of Grifo’s patients who desperately wanted a biologically related child. Willing to gamble on this coveted goal, she gave Grifo half a million dollars over 10 years to work on the technique.

It was a basic human desire combined with unfortunate circumstances, but also with the extraordinary potential that scientific research seemed to offer. These are common ingredients in questions of bioethics. Health science research is driven by many kinds of desires and is often coupled with a sense of urgency. Previously unimagined techniques seem to put distant hopes suddenly within reach.

Complexities arise

Grifo first conducted a series of experiments in mice. Once he had perfected the technique of nuclear transfer between eggs, he wanted to see if the eggs could produce viable offspring. His team implanted the eggs in mice. It worked. Several litters of healthy baby mice were born.

The time felt right to try the technique in humans. Grifo and his team had become adept at the precise and fastidious technique of nuclear transfer, and his patient, having waited while the technique was developed and perfected, was getting older. The team tried the experiment in five patients, including the woman who had funded the research.

It failed. “The eggs made with nuclear transfer fertilized and made embryos, but no one got pregnant,” Grifo explained. The eggs, it seemed, were too immature.

At New York University, Grifo is the director of the Division of Reproductive Endocrinology, the director of the Fertility Center Program, and a professor of Obstetrics and Gynecology. His line of work meets a real demand. According to the Centers for Disease Control and Prevention, nearly 7.4 million U.S. women between the ages of 15 and 44, or roughly 12% of this demographic group, have sought treatment or services for infertility. Behind these statistics are individuals and families struggling with difficult news and asking about what new treatments might become available. Although most women lack the wealth and willingness to go to such extreme lengths as Grifo’s patient did, infertility evokes deep human emotions, desires, and hopes. It also brings out deep fears.

So do some new scientific procedures, especially when they relate to creating, sustaining, or ending human life. And here bioethics gets complicated. Here, profound individual experiences of hope, desire, and fear meet with disparate societal hopes and fears, ethical questions, and a fair measure of the unknown.

To many people, bioethics sounds like an abstract idea, something official panels and committees discuss. But bioethical problems start with a story, or usually many stories, often about people having hope despite long odds. Hope to overcome a disease, to conceive, to heal from an injury. And when that story has conflict, as all good stories do, the conflict often comes in the form of fear: fear of the unknown, fear of cultural change, fear of technology, fear of ethical or moral slippery slopes.

Grifo and his team ran headfirst into that fear. One day in 2001, Grifo received a call from Susan Blumenthal, who was then the U.S. Assistant Surgeon General.

“I’ll tell you exactly what she said,” Grifo recalls: “‘What the hell do you think you’re doing up there?’ So I explained the history, the fact that we had IRB approval for all aspects of it. And she said, ‘You need to do this in monkeys first.’ Well, monkey IVF is way behind human IVF, and I don’t have any monkeys who want it.”

IRB approval—approval by an institutional review board—is a cornerstone of ethical and responsible research. Before any research can be done on animals or humans, the institution (New York University in this case) –must conduct a review of the proposed research to ensure that it conforms to ethical standards. Grifo’s research had received such approval every step of the way.

A week after the telephone call, Grifo received a letter from the U.S. Food and Drug Administration (FDA) telling him that he had to file a new drug application. He was shocked. “We weren’t doing drug research,” he says. “The FDA doesn’t regulate this kind of research. They dared me to keep doing it.” In fact, in 2001, the FDA did claim jurisdiction over nuclear transfer research. It had become clear that Grifo would have a hard time continuing this research in the United States.

Here, bioethics gets even more complicated: Science and bioethics are globalizing. Researchers collaborate across universities, countries, continents, and cultures. Worldwide, people such as Grifo’s patient face health challenges and raise hopes that drive research. Lawmakers in different countries are making different decisions about the ethics of such research. As research travels, it runs into different ethical and legal boundaries and also potentially transgresses or circumvents those boundaries. This can feed xenophobic stereotypes in which some countries are depicted as overly permissive, as fundamentally unethical. But it has been amply demonstrated that stereotypes often obscure more than they reveal.

Scene shift

At the time of Grifo’s telephone call from the FDA, John Zhang, now a well-known IVF physician in New York, was a senior research fellow in training with Grifo. Zhang had colleagues in China, and Grifo and Zhang decided to offer the researchers in China the chance to continue the work. None of them anticipated what would happen next.

On October 14, 2003, major media outlets—including The New York Times and The Washington Post–reported that a research team at Sun Yat-sen University in Guangzhou, China, had successfully impregnated a woman using eggs made by nuclear transfer. This was the team, led by Guanglun Zhuang, to which Grifo and Zhang had given their research. Although no baby was born–the three fetuses that developed from implanted eggs were delivered too early to thrive–the research nonetheless suggested that the technique was sound. Grifo recounts that the lack of success was due to obstetrical problems rather than problems with the procedure itself. “It worked,” he says emphatically.

The media focused on several concerns. One was that since human eggs contain a small energy center called a mitochondrion, which exists outside the nucleus and has its own tiny amount of DNA inherited solely from the woman who produced the egg, children born of this technique could be said to have three genetic parents: the egg donor, the woman who carried the implanted egg to term, and the man whose sperm was used to fertilize the egg. Also, concerns were expressed that research using the nuclear transfer technique was a step toward genetic engineering of human beings and human cloning. A third concern was that the technique, still experimental, might pose unknown risks to the safety of the mother and any children. Media reports highlighted the newness and riskiness of the technique, framing it as a story of questionable scientists and questionable ethics. They asked, ought we to do this kind of work? Is it too risky?

Grifo was shocked at the emergent controversy. For him, the media reports fueled public outrage and misunderstanding. He is adamant that the procedure does not constitute cloning. “Cloning is making a copy of a human being who already exists,” he said in a 2003 interview with The New York Times. “This is nuclear transfer, one element of cloning. It allows a couple to have their genetic baby, not a clone. They shouldn’t even be discussed in the same sentence.”

It is important here to clarify the distinction between reproductive and therapeutic cloning. Mention of human cloning tends to evoke the image of an identical person, but this has not yet been shown to be possible in humans. Therapeutic cloning, which is the aim of most human embryonic stem cell research, involves the production of an embryo with identical DNA to the patient from which stem cells are then harvested and used–hopefully–to treat the patient’s condition without risk of an immune reaction. Reproductive cloning has the intention of creating a genetically identical human being and is banned in most countries. Thus the debate about the use of human embryos for stem cell research involves therapeutic cloning but not reproductive cloning, even though they share techniques.

To Grifo, the issues raised in the mainstream press represented a misunderstanding of the science, the kind of misunderstanding that is often at the center of bioethical debate. The researchers saw their work as straightforward and in the interest of patients. But other people had more visceral reactions, along with complex questions about how, when, and under what conditions scientists ought to intervene, for instance, in matters such as human reproduction.

It is also hard to separate politics, economics, and culture from the controversy. Individual experiences, cultural elements, national politics, economic competition, and global politics all shape bioethics together, and each of those is somewhat influenced by, and also influences, media portrayals.

Viewing events in retrospect, Grifo says he would never have published anything until the technique produced a baby. He knows it makes a difference when and how people hear about a technique in the media. Recalling the first attempts in the 1970s to produce a baby via IVF, he notes that the first was an ectopic pregnancy and the second a miscarriage. If the press had reported on these results in today’s environment, he reasons, government regulators would have stepped in and researchers would not have been allowed to make the progress that they have in IVF. By now, more than 3 million babies have been born through IVF.

Traveling science, traveling bioethics

Bioethics gets even more complicated when deeply personal disruptions become entangled with national, international, or indeed global considerations. Bioethics frequently addresses questions of global significance that consider human flourishing and risk on a grand scale. But the experiences that it draws and deliberates on are often, at their core, deeply personal: bearing a child, watching a loved one suffer, living with a devastating disease, facing death.

The nuclear transfer experiment was, at its core, about real women with all the personal challenges that go along with pregnancy and infertility. It was a familiar story: A woman wanted to have a baby of her own and had fertility problems. She wanted the baby to be genetically related to her, not to the egg donor. This mattered to her personally, not as an abstract and theoretical question of ethics. It also mattered to the women who participated in the China study.

To Grifo, the research is ethical in that it answers to a serious problem; it is the regulations that are not ethical. Of the patients struggling with fertility problems, “I sit here and listen to them weep,” he told The Wall Street Journal. “That is powerful. And not one person writing the laws understands that.”

For Grifo and Zhuang, the tears and hopes made transnational partnership worthwhile. But if the media storm that followed was hard to foresee, so perhaps were the stereotypes embedded in that storm.

Wild East?

Reproductive biomedical research is not just about the ethics of conception; it is also about the ethics of misconception. When the West generates stereotypes about Asia, there are personal repercussions for Asian researchers, for the global research community and its supporters, and for people wanting to bear children and manage disease. The way people in the United States perceive Asia has implications for the future, for Asia, for the United States, for science, and for questions of global bioethics.

How do national boundaries matter as scientific research becomes increasingly global? As the Grifo case illustrated, it is hard enough within a single country to agree on bioethical questions. As researchers and research increasingly cross national boundaries, and because biological research increasingly has implications for all of humanity, people are asking questions about how it might be possible to establish international standards of bioethics in light of cultural differences and scientific competition.

For some people, the Grifo-Zhuang experiment smacked of ethical outsourcing. It gave rise to fears that Asia was like a new Wild West–or Wild East–of unfettered, unethical scientific practices. A Wall Street Journal report in 2003 on the work pointed to the light enforcement of regulations governing fertility clinics in China, “making China a growing haven for freewheeling research into reproductive medicine and cutting-edge genetics.” Jeffrey Kahn, the director of the Center for Bioethics at the University of Minnesota, said in a 2003 New York Times interview that he sees this kind of transnational collaboration as a way of skirting ethical issues and regulations, “as an end run around oversight and restrictions within the United States.” To the extent that bioethics is shaped by hope and fear, this is the face of the fear: fear of the unknown and often a xenophobic fear.

News representations of Chinese biotechnology at that time reflected such fear. Reports said that Chinese biologists had engaged in human cloning, that embryologists had transferred human cell nuclei into rabbit eggs, and that relatively little public debate was taking place. Such reports fueled an impassioned and fearful response in the United States. As a typical example, the New Atlantis journal ran an article in 2003 titled “Chinese Bioethics? ‘Voluntary’ Eugenics and the Prospects for Reform.” The authors referred to recent experiments in China that “raise yet more troubling questions about the ethics of biotechnology in that still authoritarian country,” and they concluded that “it is therefore a distinct possibility that the Chinese government will permit and perhaps secretly encourage the creation of cloned or genetically modified children for the ‘good of society.’”

Such research projects do indeed merit serious attention. They should provoke intense scrutiny and ongoing public and governmental consideration wherever they are conducted. Although many of these same kinds of research projects were underway at Western sites, in the media these researchers were mainly characterized as rogue scientists who were seeking fame and fortune, or as marginalized “sects” or “cults”—in other words, as individuals rather than representatives of a country. But in discussions of Asia, and of China in particular, questions of bioethics were framed at the level of a people, culture, country, or region.

Bioethical institutions were developing in China even as these controversies were taking place. Ole Doering, a China specialist and philosopher, reports that a “new wave of infrastructure building to regulate and monitor biomedical activities in China took off in 1998.” He, too, writes about ethical outsourcing, but from a different angle. Doering quotes a semi-official Chinese daily newspaper warning in 2003 that “we must be aware that some scientists from developed countries make use of the ignorance and eagerness of their colleagues in the developing countries to carry out experiments banned in their own nations.” In this view, ethical resistance to the Grifo collaboration from a Chinese perspective might not so much question unscrupulous Chinese researchers, but unscrupulous and exploitative foreign, and implicitly Western, collaborators.

Racing ahead

The implications of the Grifo-Zhuang nuclear transfer aftermath reached far beyond fertility treatments and reproduction. The nuclear transfer technique was also seen as central to the promise of stem cell research. The hope was that if one could replace the nucleus of a human embryo with a nucleus from a patient’s cell, then get it to develop for about a week or so, one could get stem cells with that patient’s DNA. This meant that these stem cells could be used to potentially regenerate almost any kind of damaged tissue without prompting an immune response. The potential to treat formerly intractable conditions seemed close at hand.

Nuclear transfer thus holds high stakes and high potential in stem cell research. And stem cell research is frequently characterized as a race: among competing scientists, laboratories, and countries, as well as for cures, money, and fame. The phrase “stem cell race” abounds in the press. The phrase’s popularity was fueled, in part, by the restrictions that President George W. Bush placed in 2001 on federal funding for human embryonic stem cell research. The restrictions limited federal funding to a few existing human embryonic stem cell lines, the so-called presidential lines.

Scientists quickly expressed concerns that the restrictions would threaten this field of medical science. Patients and families worried that treatments and cures would be delayed. Politicians and venture capitalists worried that their regions and investments would be hurt by restricted research funding. Fears and hopes continue to cycle through these bioethical debates.

Scientific globalization evokes images of international competition, every country trying to get ahead. Yet while many Westerners worried that Asian countries would race ahead, unfettered by research and ethical regulations, the inverse may actually be happening in some places. In countries where no regulations yet cover such practices as nuclear transfer or stem cell research, some researchers feel reluctant or even afraid to work in controversial fields without a green light from policymakers, ethicists, and the public. Indeed, policymakers in many countries are working hard to develop ethical research guidelines. Although some people still think of regulations as stifling research, a lack of formal guidelines could be worse, if this means that researchers are not certain what is culturally or legally permissible, now or later.

In China, regulators moved quickly in the aftermath of the Grifo-Zhuang nuclear transfer pregnancy story to ban the procedure. Despite China’s quick response, stereotypes persist, as pointed out by Erica Jonlin, clinical research administrator and regulatory manager at the University of Washington Department of Medicine, whose daily work involves questions of ethics, research, and stem cells. On one hand, she says, “Scientists can collaborate. Scientists like to collaborate.” But she says there remains a stereotype in the U.S. scientific community that scientists in China can do anything. In fact, if there ever were regulatory advantages to doing research in China, they’ve largely gone away. But the fear of unfettered Asian research continues.

Experiences in Taiwan

In some ways, China acted as a stand-in for a broader U.S. cultural fear about Asia, and East Asia in particular. Indeed, many people in Asian countries did see stem cell research, and biotechnology more generally, as a new hope, a way to catch up with the West on the global stage of science. They also saw it as a way to bolster their economies. Singapore developed Biopolis, a state-of-the-art biotech site based on the model of Silicon Valley and famous for recruiting highlevel scientists from the West. South Korea developed a wellfunded stem cell research laboratory at Seoul National University and seemed poised to become a global leader in stem cell research until a scandal involving its leader, Hwang Woo-suk, broke in late 2005.

Taiwan, too, announced in 2005 a national project to develop the country as “Biomedtech Island”–an Asian hub for biomedical technology. Underscoring the urgency, a minister of Taiwan’s Science and Technology Advisory Group said in a 2005 report in Taiwan News, “We are under pressure of time to get the ‘Taiwan–Biomedtech Island’ plan going as soon as possible.” Pointing out similar projects underway in China and Singapore, the official said Taiwan hoped to “compete well in the advanced biomedical fields and become the leader in the field in Asia.” Stem cell research was an important part of this plan.

At Academia Sinica, Taiwan’s most prestigious research institution, broad open spaces and rows of palm trees frame the state-of-the art science facilities. There, until recently, John Yu headed the stem cell research program. He and his wife, Alice Yu, left successful scientific careers in San Diego to help build Taiwan’s biotech sector.

In practice, John Yu spends much of his time not at the laboratory bench but on the development of ethical research protocols. He founded the Taiwan Society for Stem Cell Research, which developed a scientific network and holds discussions on how best to regulate research. He served as Taiwan’s representative at the International Society for Stem Cell Research and was a member of the task force that developed in 2006 the society’s “Guidelines for the Conduct of Human Embryonic Stem Cell Research,” a global standard for ethical stem cell practice.

He is a vocal critic of unregulated stem cell research and therapeutics. Work in Taiwan and elsewhere that is perceived as unethical risks resulting in public opprobrium not only for the individual researcher or physician, but also for the science itself. And although many people in the United States worried that an Asian lack of regulation and ethical constraint would create an atmosphere of unfettered and unethical research, in Taiwan, the opposite seemed to occur.

Consider the case of one young Taiwanese stem cell scientist. (Given the sensitivity of his position, he would rather not be named, so he will be called Dr. Li.) Beneath his softspoken and unassuming demeanor, Dr. Li exudes a passion for his work. For him, stem cell research has both deeply personal and national stakes. He grew up in Taiwan, then completed his education and training as a stem cell biologist in the United Kingdom and the United States. He began a family and was developing a promising career in the United States when he returned to Taiwan in 2004. Like John and Alice Yu, Dr. Li returned to help build biotech in his home country. “Maybe this will sound naïve,” he says, “but originally I came back to Taiwan because I had this idea that it’s my duty; that maybe I can help Taiwan a little bit on stem cell research.”

In his previous work, Dr. Li had used dozens, perhaps hundreds, of human embryos. But in Taiwan, by 2007, when guidelines were still waiting for government authorization, he had not used a single one.

Instead, he helped to establish such guidelines and found himself in deep reflection about the ethicality of his own research. Rather than speeding up his research, the lack of clear policy in Taiwan slowed it down. It seemed that the established policies in the United Kingdom and the United States had enabled him to focus on his research, shielding him perhaps from deep ethical reflection of the type that now holds his attention. He also attributes this shift to more personal factors, such as his maturation and the birth of his first child. He recognized the potentiality that inheres in the human embryo. No longer seeing an embryo as just a research object, he came to see its potential, given just the right set of extremely contingent circumstances, to become someone’s child. He understands the hopes that stem cell research inspires, and the fears, too.

Stem cell research has numerous risks. Individual scientists risk their reputations, careers, and even their freedom if they conduct work that is deemed unethical. Treatments are risky for patients. The science itself relies on public support. Researchers worry that hype and premature human treatment might ultimately diminish support. John Yu of Academia Sinica says this is the greatest worry for stem cell researchers: “We don’t want society to expect too much in terms of what we can achieve now.” He cites a U.S. survey which suggests that the general public’s expectations about therapy developments from stem cells are much more optimistic than those of stem cell scientists. His concern is that hype and “unregulated” physicians will lead the public to expect too much too soon, thus setting the stage for the fragile support of stem cell research to be undermined when therapeutic production is slower.

Dr. Li is also concerned about public attitudes toward stem cell research, saying that “everybody that works in this field, they really want to know what is the public opinion.” It is personal for him: “Myself, I want to know. I really want to know, what do they think about this.” Although stem cell research is still not a major topic of public conversation in Taiwan, some insights about public attitudes may emerge from a study led by Shui-chuen Lee, a Confucian philosopher and bioethicist, and Duujian Tsai, a sociologist and community organizer. A team led by these two professors conducted surveys to identify public knowledge and public concerns about stem cell research.

As Dr. Li and John Yu know, it is not enough to progress scientifically; science has to be done carefully and correctly at every step. Taiwan became a full electoral democracy in 1996, after a 12-year transition period. Before, Taiwan was ruled under martial law for 38 years. So, domestically, public inclusion has become an important topic of governance– political and scientific. And internationally, public inclusion has become an important component of responsible scientific decisionmaking. It is not enough to have bioethical and research policies; increasingly, such policies have to both represent broad public consensus and conform to international standards.

The California experience

While people in Taiwan were taking surveys to assess their knowledge of and support for stem cell research, across the Pacific, Californians were showing their support at the ballot box. In a heavily funded campaign, proponents of Proposition 71, the California Stem Cell Research and Cures Initiative, asked the public to support stem cell research.

The campaign was successful. Passed in 2004, the initiative mandated state investment in stem cell research: $3 billion over 10 years. Proposition 71 represented a new kind of public engagement with science. With federal funding for most human embryonic stem cell research halted in 2001, California and several other states, including Connecticut, Illinois, New Jersey, New York, and Maryland, subsequently took it upon themselves to fund this type of research.

In California, many supporters saw a vote for Proposition 71 as a hopeful vote against President Bush and what they perceived as an anti-science, ideologically driven, and fear-building regime. They saw their vote as progressive, pro-science, and pro-cures, with real people’s lives at stake. They also saw it as a more democratic, if risky, way to fund science. In California, public input has come to be seen as a necessary element in doing ethical science, in both research and its funding.

In a sense, then, bioethics has become explicitly context-specific. Classically, the field of ethics poses such questions as “What should I to do?” and “What constitutes the good life?” But when people encounter fast-changing biomedical technologies, these questions can be especially difficult to answer.

For Jeff Sheehy, advocacy is the answer, as witnessed by his active involvement in various aspects of HIV/AIDS work. He successfully established organ transplantation programs for people living with HIV in California and nationally. He is open about his own struggles in living with HIV. At a 2010 meeting of the California Institute for Regenerative Medicine (CIRM), he said, “For instance, I’m 53, so I’m here”—pointing to a graph of life expectancy for those living with HIV/AIDS—“and it’s a real bet for me whether I’m going to make my five-year-old daughter’s wedding, unless …”

In November 2004, Sheehy received a call from the leader of the California State Senate, John Burton, asking him to accept an appointment as a patient advocate to the governing board of CIRM, which was established by the passage of Proposition 71. Still on the board, he is also now director for communications at the University of California, San Francisco AIDS Research Institute.

Uniquely, CIRM included a mandate to include the state’s diverse communities in every aspect of its decisionmaking process. As a result, these communities help in addressing a range of issues, such as determining which supply companies to use and setting mandates for preferential pricing for the state on any procedures and products to emerge from CIRM-funded research. Proposition 71 was seen to hold more than just the potential to produce cures for various medical conditions; it was seen as a way for the state to gain a foothold in what held promise to become a burgeoning field of the biotech economy.

Writ more broadly, debates in the United States about stem cell research have mainly centered on human embryonic stem cell research and questions about the moral status of the human embryo. In a way unique to this country, the debates are shaped strongly by the divisive abortion issue. For some U.S. residents, the destruction of a human embryo on the research bench is equivalent to something like murder.

For Jeff Sheehy, this is a false argument. “It seems like the whole embryo argument here has been misunderstood,” he says. He says that the embryos are not created for research, but are excess embryos “created to fulfill people who wanted to have children” using IVF. It would be better, he adds, for people who oppose the research to also support public funding of IVF, thereby reducing the strong financial incentive to create as many embryos as possible with each IVF cycle. This would, after all, reduce the overall number of embryos created.

Ultimately, Sheehy suggests, his voice rising slightly, the decision of whether to destroy these excess embryos, to donate them to science, or to give them to others seeking IVF should lie with the parents. “These are ethical choices for parents,” he says. “They should have the autonomy as Americans and as parents.”

Here, Sheehy appeals to the values of anti-paternalism, individual autonomy, and parental decisionmaking that he sees as hallmarks of the United States. He also brings up some broader questions: What kinds of matters should be private and what kinds should be public—and what kinds of things should be publicly funded? The ongoing health care debates and the recession have revealed much about the deep divisions that exist about such topics.

For countries, scientists, and patients, the stem cell race is afoot. Each group experiences a sense of urgency, but none more so than those of patients waiting and hoping for treatments and cures. Sheehy feels this acutely; he has seen his community devastated by HIV/AIDS and, after all, he hopes to make it to his daughter’s wedding.

It is turning out that stem cells look like they may be able to cure HIV infection. In 2008, doctors in Berlin reported that a stem cell transplant had functionally cured a patient with HIV. In 2009, CIRM committed up to $20 million for a study to replicate the results. This would not have happened without Sheehy on the governing board. Many of the board’s members thought that such a sizeable investment was unnecessary. After all, HIV/AIDS in California is being relatively well managed by combinations of antiretroviral drugs.

Sheehy argued, however, that these drugs are problematic and that he and many of his friends would happily trade them for the hope offered by a stem cell therapy. He described for the board the significant side effects of these medications and recounted in personal terms the increased rates of heart disease, non–HIV-related cancers, and neurological deficits that accompany HIV/AIDS infection. When critics discourage funding for stem cell therapies because they do not think anyone with HIV will participate in a clinical trial of such experimental procedures, he is there to say, “I would.”

Changing landscape

The stories of Jeff Sheehy’s activism, public dialogues in Taiwan, and James Grifo’s patient all suggest that the relationship between the scientific sphere and the public sphere is changing. No longer are scientists seen as appropriately selfregulating. CIRM’s inclusion of 10 patient advocates on its governing board also signals a new way of funding and guiding science.

It is also becoming increasingly clear that context matters—cultural, geographic, economic contexts surely, but also the specific details of each case. The mainstream media framed the Grifo-Zhuang case as controversial science, but it left out the context in which a woman, desperate for a biologically related child, prompted and funded the research. Although this detail may well raise additional questions about the ethicality of such funding arrangements, the details nonetheless matter. Individual patients are shaping emerging research.

On the international stage, despite variations in how different countries approach bioethics, the guidelines for human embryonic stem cell research developed by the International Society for Stem Cell Research have found relative acceptance in almost all countries where such research is being conducted. Also, in 2011, nearly a decade after the Grifo-Zhuang controversy, Britain’s esteemed Nuffield Council on Bioethics approved a new IVF technique that involves replacement of the mitochondrion rather than the entire nucleus of a patient’s egg. Though this approach raises very similar ethical concerns, the media response to date has been fairly neutral.

Slowly changing mores are not comforting to someone hoping for a cure to a disease or a chance to bear a child. Nor are they comforting to people who see them as a slippery slope that threatens human integrity and flourishing. But increasingly, locally and globally, bioethical decisions are including more voices, of individual scientists and patients and activists alongside scientific leaders and formal ethicists. Science and bioethics are indeed global endeavors, and now new kinds of relationships and new voices are emerging within and across borders.

From the Hill – Fall 2012

R&D funding picture remains mixed, as budget negotiations stall

The laborious process of crafting a federal budget for the next fiscal year (FY) 2013 appeared set to grind to a halt, when Senate Majority Leader Harry Reid (D-NV) and House Speaker John Boehner (R-OH) said on July 31 that they had reached an agreement on a continuing resolution to fund the government through March 2013. The agreement became necessary when congressional leaders recognized that with the election looming, neither chamber was likely to approve the 12 spending bills before the beginning of FY 2013 on October 1.

The agreement also appeared to settle a running dispute between the parties over total FY 2013 spending. The 2011 debt-ceiling agreement established a discretionary spending cap of $1.047 trillion. Although the administration and Senate Democrats have abided by this agreed-on limit, the House GOP passed a budget resolution that capped overall spending at $1.028 trillion. The lower cap had drawn the ire of Democrats, and the White House had consistently promised to veto any spending bill that abides by the lower cap. The House position has been jettisoned, at least for now.

Despite the apparent settlement of the dispute, Congress remains at an impasse on negotiations to avert the “sequestration” cutbacks required for both defense and nondefense spending set to begin in January 2013. Negotiations are under way to avert the across-the-board 10% cuts to defense and 8% cuts to nondefense spending, and a number of Republicans have said they may be willing to consider revenue increases as a part of the package. Even so, support for deep cuts remains strong in some quarters, particularly for nondefense discretionary spending, a category that includes virtually all federal spending outside of defense and entitlement spending and virtually all nondefense R&D. Nondefense discretionary spending was targeted for cuts in the House-passed budget resolution, and similar proposals to protect defense spending at the expense of nondefense spending have been attached as riders to other bills.

To combat these attempted cuts, approximately 3,000 organizations from across the public interest spectrum recently sent a letter to Congress asking for a responsible deficit-reduction approach that does not include further cuts to nondefense discretionary spending, which has already been cut by about 10% since FY 2010. Projecting the effects of this approach into the future, the American Association for the Advancement of Science (AAAS) has estimated that shifting the planned cuts entirely to nondefense areas could result in a reduction of 18%, or $52 billion, in nondefense R&D funding at science agencies over the next five years.

Pressure also remains intense on the defense side, as defense contractors, who have long argued that the cuts will force them to fire thousands of employees, have said they will be required under federal law to issue layoff notices by November 2, which is 60 days before the cuts begin to take effect and just a few days before the elections. Reports by the National Association of Manufacturers and the Aerospace Industry Association have placed budgetcut– induced job losses at one to two million. Congress passed legislation that requires the administration and Pentagon to explain exactly how the across-the-board cuts would be allocated, as the Office of Management and Budget begins consultations with the agencies on these questions.

Even as the overall picture remains cloudy, there nevertheless has been some progress on spending bills on the House side. On August 2, the Senate Appropriations Committee passed its version of the FY 2013 Defense Appropriations bill, after the full House passed its own version two weeks earlier. According to AAAS estimates, the bill would reduce Department of Defense (DOD) R&D by $1.9 billion or 2.5% below FY 2012 levels, nearly equal to the overall cut proposed by the administration. In contrast, the House version would reduce overall DOD R&D by about half as much. Basic research across all military departments and agencies would be funded at roughly $2.1 billion under all three proposals, whereas applied research funding in the Senate bill is more generous than in either the administration or the House proposals. As in prior years, and mirroring the House, the Senate Committee has restored substantial funding to R&D in the Defense Health Program, which the administration had targeted for a nearly 46.9% cut. The FY 2013 Interior/Environment spending bill is now the lone R&D-heavy spending legislation yet to be taken up by the Senate Appropriations Committee.

A House appropriations subcommittee passed the Labor, Health, and Human Services spending bill on July 18, which would keep National Institutes of Health (NIH) funding flat for FY 2013, similar to the Senate version of the bill. The House bill would terminate the Agency for Healthcare Research and Quality, and mandate a 90/10 split for extramural/intramural research and a 55% split for basic research at NIH, in an attempt at ensuring that both kinds of research remain priorities for the agency. The subcommittee would also reduce the maximum salaries that institutions can charge to NIH as a cost-saving measure.

On June 29, the full House voted 261 to 163 to approve the FY 2013 Transportation and Housing and Urban Development spending bill. According to AAAS estimates, the Department of Transportation would receive approximately $1 billion for R&D funding in FY 2013, an increase of $70 million or 7.4% above FY 2012 levels, although less than the president’s request. The Federal Highway Administration would receive most of the R&D boost sought by the administration, reaching $494 million, 20.2% higher than in FY 2012.

On June 28, the House Appropriations Committee approved its FY 2013 Interior and Environment appropriations bill, which would slash R&D funding at the Department of Interior, the Environmental Protection Agency (EPA), and the Forest Service. According to AAAS estimates, the bill would fund Interior R&D at approximately $740 million, $122 million or 14.2% below the president’s budget request and $56 million or 7.1% below FY 2012 levels. U.S. Geological Survey (USGS) R&D would be cut by 14.4% below the president’s request and 8% below FY 2012. EPA funding would be reduced by 10.2% below the president’s request and 8.9% below FY 2012. The cut is almost entirely in EPA science and technology, and the committee also passed several amendments to limit the EPA’s ability to regulate greenhouse gas (GHG) emissions and toxins.

On June 19, the House Appropriations Committee approved its FY 2013 agriculture funding bill (H.R. 5973). According to AAAS estimates, the bill would cut U.S. Department of Agriculture R&D by 4.5% below the president’s request and 5.9% below FY 2012, although part of this apparent cut is attributable to the end of the biomass R&D program, which is up for reauthorization in FY 2013 in the current farm bill. Although the Agricultural Research Service and the National Institute of Food and Agriculture would be cut, the Agriculture and Food Research Initiative would receive a boost of 4.6% above current-year funding. The bill generally falls short of the president’s request and the current Senate version of the bill (S. 2375), which was passed by committee on April 26 and awaits floor action.

House Republicans hold controversial hearings on EPA rules

The House held two controversial hearings on the EPA on June 6, addressing the effects of recent rules on the oil and gas industries. The Committee on Energy and Commerce’s Subcommittee on Energy and Power invited several stakeholders to express their concerns about EPA enforcement in Region 6, which includes Arkansas, Louisiana, New Mexico, Oklahoma, and Texas. The House Science, Space, and Technology Committee’s Subcommittee on Energy and Environment’s hearing focused on the costs and benefits of recent EPA rules, featuring witnesses from industry groups.

House Republicans used the hearings to criticize the EPA, saying it interferes with state regulations, unfairly burdens coal and other fossil fuel industries, and creates standards based on faulty science. At the Energy Committee hearing, Barry Smitherton, chairman of the Texas Railroad Com- mission, discussed the Range Resources case. In December 2010, the EPA issued an emergency endangerment order to the Range Resources Company against the advice of the commission, which serves as a state regulatory agency. It was later discovered that Range Resources was not responsible for the groundwater pollution that the EPA had detected, and the order was lifted, but only after the company spent millions of dollars defending itself, according to Smitherton.

At the Science Committee hearing, Tom Wolf, executive director of the Illinois Chamber of Commerce Energy Council, said the New Source Performance Standards for carbon dioxide emissions worked for natural gas plants but were impossible for coal-powered plants to meet with the currently available technology. He said the large leap in standards would create a roadblock for coal producers, instead of the intended incentive for innovation.

At the same hearing, Energy and Environment Subcommittee Chairman Andy Harris (R-MD), and Michael Honeycutt, chief toxicologist at the Texas Commission on Environmental Quality, discussed what they saw as the EPA’s overestimation of the benefits conferred by its rules.

Democrats on both subcommittees criticized the hearings’ intentions. Energy and Environment Subcommittee Ranking Member Brad Miller (D-NC) called the Science Committee’s hearing “one more forum for specific big industries to air their grievances about the EPA.” Energy Committee Ranking Member Henry Waxman (D-CA) called his colleagues’ opening statements “part of the fact-free, anti-EPA rhetoric of the Republicans.”

Bills introduced to improve forensic science

Kirk Odom served 20 years in prison for a crime he did not commit. He was convicted on the basis of a mistaken victim identification and faulty forensics. Thirty years later, DNA testing on a hair found at the crime scene, as well as stains on pillowcases and the victim’s clothing, proved that he was innocent. Odom was the third person in three years to have his conviction overturned because of unreliable hair analyses in Washington, DC.

Nationwide, there have been 292 post-conviction DNA exonerations in the United States since 1989, and, according to the Innocence Project, a nonprofit that helps prisoners who were wrongly convicted, about half of those wrongful convictions were due at least in part to poor forensic science.

Congress is considering several new bills in response to these recent events, as well as an investigative report by the Washington Post and a 2009 National Research Council (NRC) study, Strengthening Forensic Science in the United States: A Path Forward.

Last year, Sen. Patrick Leahy (DVT) introduced the Criminal Justice and Forensic Science Reform Act (S. 132). It would establish an Office of Forensic Science in the Department of Justice (DOJ) that would be responsible for creating and implementing uniform standards and enforcing regulations, as well as a Forensic Science Board to determine research priorities. (The placement of such an office in the DOJ runs counter to the NRC report, which said that a new and independent organization would be needed to regulate the forensic community, because no existing government agency has a relevant mission statement or the appropriate resources to take on this task.) The bill would also require that any labs or individuals receiving federal funding be accredited based on standards outlined by the new Board and Office of Forensic Science. The bill is currently being considered by the Senate Committee on the Judiciary.

In July, Sen. John D. Rockefeller IV (D-WV) introduced the Forensic Science and Standards Act of 2012 (S. 3378), which directs the National Institute for Standards and Technology (NIST) to develop standards for forensic scientists and establish a Forensic Science Advisory Committee. It would be composed of research scientists, forensic scientists, and members of the legal and law enforcement communities and chaired by the director of NIST and the attorney general. Rockefeller’s bill would also establish a National Forensic Science Coordinating Office in the National Science Foundation to develop a research strategy and provide grant money for forensic science research centers. S. 3378 is currently being reviewed by the Senate Committee on Commerce, Science, and Transportation. Rep. Eddie Bernice Johnson (D-TX) introduced a companion bill, H.R. 6106, which has been referred to the House Committees on Science, Space, and Technology, as well as the Judiciary Committee.

The proponents of the bills believe that nationally recognized standards and a strong peer-review process, much like the one that helps to regulate the rest of the scientific community, will result in better research and more accurate analyses and lead to fewer wrongful convictions.

Senate committee examines EPA rule on air pollution from fracking

The Senate Committee on Environment and Public Works’ Subcommittee on Clean Air and Nuclear Safety held a hearing on June 19 to review new EPA air standards for hydraulically fractured natural gas wells and oil and natural gas storage.

The EPA rule, which was finalized on April 18, includes the first-ever national standards on air pollution from gas produced in wells using a process known as fracking. Members of the oil and gas industry have criticized the new rule for its potential impact on domestic natural gas production.

Opening statements and comments at the hearing fell along party lines. Subcommittee Chairman Thomas Carper (D-DE) praised the EPA for addressing the lack of fracking regulations in most states, and Sen. Benjamin Cardin (D-MD) insisted that air pollution is a national issue because it does not follow state boundaries. On the other side, Subcommittee Ranking Member John Barrasso (R-WY) criticized the Obama administration for working against the natural gas industry, despite its “all-of-the-above” energy rhetoric. Committee Ranking Member James Inhofe (R-OK) brought up recent EPA controversies and argued that the regulation of fractured wells should be left to the states.

The first panel featured Gina McCarthy, the EPA’s assistant administrator for the Office of Air and Radiation, who said the recent rule on air emissions was achievable, would result in cost savings, would reduce air pollution, and would not slow natural gas production. She also described changes made to the final version of the rule in response to industry feedback, which included the introduction of a transition period before the use of reduced emission completions (also known as “green completions”) would be required. Green completions capture natural gas that is emitted during a well’s flowback period, preventing the release of volatile organic compounds into the atmosphere. The final rule also includes a new subcategory of wells in low-pressure areas. These wells are not required to use green completions, because the new technology is not cost-effective in those cases.

On the second panel, Fred Krupp, a member of the Secretary of Energy’s Advisory Board Natural Gas Subcommittee, outlined the subcommittee’s findings that oil and natural gas production results in the emission of toxic air pollutants such as carcinogenic benzene, ground-level ozone, and methane, which he said causes global warming at a rate 72 times higher than that of carbon dioxide.

John Corra and William Allison, representatives of state regulatory agencies in Wyoming and Colorado, respectively, highlighted the current environmental regulations in their states, which the EPA used as the basis for the new rule. Both stressed the importance of allowing states flexibility during implementation, because the use of green completions is not technologically or economically feasible at some sites.

Tisha Schuller, the president and chief executive officer of the Colorado Oil and Gas Association, expressed her concerns that the EPA overestimated the benefits of the rule by overestimating the emissions from fractured wells and overestimating the cost savings from the rule, while underestimating the costs for new equipment and regulatory and administrative requirements. Darren Smith, the environmental manager at Devon Energy Corporation, suggested that the EPA overestimated the current emissions from fractured wells.

Carper ended the hearing by hailing the rule as “common sense.” Although the witnesses agreed that the rule needs some tweaks before it is fully implemented in 2015, the debate overall seemed to be, as McCarthy said, a question of whether the rule was “good or very good.”

Federal science and technology in brief

  • The AAAS Office of Government Relations has developed a Web site (http://elections.aaas.org/2012/) that describes and tracks the presidential candidates’ positions on science, technology, and innovation issues.
  • New rules proposed by the Small Business Administration for the Small Business Innovation Research program and the Small Business Technology Transfer program are causing a stir among industry advocates. The rules could eliminate a requirement that grant applicants be majority-owned by a U.S. resident or company. Instead, they propose that applicants must operate primarily in the United States or “make a significant contribution to the U.S. economy,” without a U.S. ownership requirement. There are also concerns that the new proposal could create a loophole that opens the door for large companies, in addition to small businesses, to receive grants.

“From the Hill” is adapted from the newsletter Science and Technology in Congress, published by the Office of Government Relations of the American Association for the Advancement of Science (www.aaas.org) in Washington, DC.

Forum – Fall 2012

The U.S.-Mexico connection

Christopher Wilson’s “U.S. Competitiveness: The Mexican Connection,” (Issues, Summer 2012) seeks to open our minds to an alternative way to think about competitiveness. Too much of the debate on competitiveness has swung between those who stress deregulation and those who advocate a form of industrial policy. Wilson suggests that an alternative path should aim at deepening integration between the United States and Mexico. He is right, but his insights would be more powerful if they were stretched to include all of North America and if we looked beyond joint manufacturing to a fundamentally different continental relationship.

Instead of thinking about Mexico as a source of grisly drug-related violence and poor illegal aliens, Wilson reminds us that Mexico is our second-largest market, the second-largest source of energy imports, the destination of about $100 billion in U.S. direct investment, and a growing source of foreign direct investment. Canada, however, is the first market for U.S. goods, and it is also the largest source of U.S. energy imports.

Our continent is the largest free-trade area in the world, but its leaders have failed to build on the early success of the North American Free Trade Agreement (NAFTA). Although the North American share of global production increased from 30 to 36% between 1994 and 2001, it has since shrunk back to where it was before 1994.

Wilson proposes that the United States and Mexico expedite plans to make the border more efficient, expand transportation links and infrastructure, dismantle rules-of-origin provisions one product at a time, and confront shared challenges in Asia and elsewhere. These are all good ideas, but they would be much better if Canada were an integral part of the strategy.

The problem is that Canada shows little interest in collaborating with Mexico, and the United States seems to prefer to approach one country and one problem at a time. Because each of the problems involves powerful domestic or bureaucratic interests that prefer the status quo, we have made almost no progress on Wilson’s agenda.

What is needed is a farsighted vision that recognizes that the best way to improve competitiveness is to create a seamless continental market. The best way to ensure American security is to turn the borders into a bridge of cooperation where officials of all three countries work from the same set of procedures. The best way to ensure that Mexico moves to the first world economically and that Canada designs new trilateral institutions is for the United States to adopt a new approach to leadership that respects its neighbors as partners in the global competition with Asia.

The only way to achieve these goals and move forward is to begin with the idea of a continent of promise. If all three countries begin to see their challenges as North American, they will find continental solutions where they previously saw only chronic national problems.

ROBERT A. PASTOR

Director, Center for North American Studies

Professor of International Relations

American University

Washington, DC

The author’s recent book is The North American Idea: A Vision of a Continental Future.


The conventional wisdom in the United States about NAFTA and Mexico has never been accurate. For years, the prevailing view in many circles has been that NAFTA cost the United States millions of jobs that relocated to Mexico seeking lower labor costs and an environment with lower health and safety standards. A more accurate description of what actually happened is that NAFTA extended the competitive life of thousands of U.S. companies and jobs by allowing them to relocate part of their production to Mexico, where labor costs were more competitive.

Christopher Wilson’s article does an excellent job of describing how the “Mexican Connection” has served U.S. economic interests. The evidence that he presents illustrates some of the many ways in which the trade and investment flows that NAFTA enabled caused the two economies to become much more closely integrated, in a relationship that is best characterized as a partnership and not a competitive rivalry.

The evidence is clear that NAFTA has been highly beneficial to both economies, improving the competitiveness of entire economic sectors and extending the life of thousands of companies that might otherwise have failed, on both sides of the border. The impact of NAFTA has been much greater than simple trade statistics suggest: It created incentives that caused entire sectors to restructure operations, leading thousands of companies on both sides of the border to specialize in activities in which their relative costs and productivity made them more competitive. The restructuring process caused Mexico’s economy to become more specialized in labor-intensive manufacturing activities (such as the assembly of autos) while providing incentives for U.S. companies to focus on high–value-added activities (such as the design and manufacture of high–value-added components). What happened as a result of NAFTA is exactly what trade theory predicted should happen.

The global economic crisis has made specialization even more valuable. To succeed in the current economic context, it is indispensable to ensure that products are highly competitive. Hence, producing in locations where labor and logistics costs are lowest has become imperative. For this reason, Wilson argues that greater attention must be paid to ensuring that the cost of transporting goods between the three NAFTA partners is optimized, but without putting the security of any of the partners at risk. He also argues that is vitally important to ensure that the NAFTA partners develop a shared view of their global trade interests, acting in concert in trade negotiations.

Wilson is absolutely right, but why stop there? His recommendations focus primarily on ensuring that the partnership continues to work effectively in support of shared production processes. But focusing exclusively on tradable commodities can cause the governments of the three countries to overlook enormous competitive opportunities in the service sectors of the three countries.

The opportunities in services are enormous. For instance, rising U.S. health costs could be contained by allowing health care providers in Mexico and Canada to compete in the United States. Currently, both countries are capable of providing high-quality health services at much lower costs than the United States.

The opportunities do not stop there: Mexico’s economy would gain competitiveness if it had access to the worldclass engineering and construction services available in the United States and Canada; and the three partners would gain competitiveness by integrating their electricity sectors or agreeing to the same technological standards in petroleum services and telecommunications. U.S. agriculture would remain competitive if a formula can be found to make inexpensive farm labor systematically available when and where it is needed.

Wilson has written a very valuable essay. His arguments are correct, but the opportunities are actually much broader than those that he points to and should be captured for the good of the three partners in the region.

ROBERTO NEWELL

Vice President

Instituto Mexicano para la Competitividad

Mexico City, Mexico


Drug policy research

In “Eight Questions for Drug Policy Research” (Issues, Summer 2012), Mark A. R. Kleiman, Jonathan P. Caulkins, Angela Hawken, and Beau Kilmer argue that more of the research on drug abuse should be guided by the goal of reducing social costs. Those costs are associated not only with the multiple impairments of those who abuse, but also with the violence associated with underground drug markets and the great burden that drug enforcement places on the criminal justice system.

Last February, the director of the National Institute on Drug Abuse (NIDA) responded to the president’s budget proposal for fiscal year 2013 with a statement that NIDA’s research priorities were to develop better medications to combat addiction, support basic research on genetics and brain functioning, and translate scientific evidence into improved clinical treatment. These goals reflect the medical model that the problem is “addiction,” which is a “disease” that can be combated through treatment or prevented through vaccines. That may be a reasonable focus, given the NIDA mission, but we are left with the question of which scientific agency is going to engage with the dominant approach of drug policy as it actually exists in the United States today. Hundreds of thousands of drug offenders, mostly dealers, are arrested, convicted, and imprisoned each year in the quest to reduce the availability of drugs. Could scientific research improve on this program, in the sense of developing a more cost-effective approach to reducing abuse and its costly consequences? The answer is surely yes, and the Kleiman et al. article provides some promising leads.

A case in point is HOPE, which the authors briefly discuss. Stunning results with “coerced abstinence” for felony convicts were reported from a randomized controlled trial in Honolulu. The U.S. Department of Justice Office of Justice Programs deserves much credit for funding a new experimental replication in four sites. The basic approach may seem to contradict the medical model, because the only mechanism by which the threat of swift and sure punishment could reduce drug use is the volition of the drug-abusing convicts. HOPE suggests that much of the drug abuse in this population could be eliminated by getting the incentives right, and at no net cost to society. If that result is replicated, it will surely inspire discomfort with the traditional medical model, while providing directly relevant guidance to policymakers.

The HOPE story has a larger moral to it. Translational research is usually understood in terms of scientific understanding, as developed in the laboratory, being put to work in shaping practice. But the apparent success of coerced abstinence is an example where the “translation” should go in the opposite direction. If brain science ignores the role of volition in drug abuse, then it is missing out on what appears (from this field experience) to be an important mechanism, and one that is particularly relevant to policy design. The translation effort should be in both directions. We need to make sure that federal funding agencies encourage this exchange.

PHILIP J. COOK

ITT/Sanford Professor of Public Policy

Sanford School of Public Policy

Duke University

Durham, North Carolina


Communicating uncertainty

In “Communicating Uncertainty: Fulfilling the Duty to Inform,” (Issues, Summer 2012), Baruch Fischhoff offers a cogent and balanced set of actions that experts and decisionmakers can take to improve the way in which uncertainty is handled when experts are asked to inform decisions. In particular, he highlights how the concepts and consequences of uncertainty, probability, and confidence are often misunderstood and sometimes conflated, and how these same problems provide insights into the many potential strategies for addressing them.

I suggest several points of clarification that both strengthen and complicate Fischhoff ’s arguments. First, policy-and decisionmaking face increased scrutiny from the public, and this scrutiny can lead scientists and decisionmakers to approach uncertainty in ways not fully addressed by Fischhoff. Skepticism about climate change is an excellent example of how the public can become aware of an issue but profoundly misunderstand the nature of uncertainty within it. The political views that this misunderstanding motivates can then drive policy and decisionmaking, even when policymakers actually have a deeper understanding than the public of how uncertainty should alter decisions, because they ultimately respond to their constituents. Scientists may also begin to present uncertainty differently in the face of this public response, essentially trying to influence decisions to be more in line with where the experts think they should be, knowing the public’s misunderstanding. Effective decisionmaking clearly requires that scientists and decisionmakers also engage the public about the realities and consequences of uncertainty.

In conservation biology, my research field, issues of uncertainty play out in concrete and visceral ways; for example, when decisions must be made about how to manage threatened or endangered species. Mishandling of uncertainty in such decisions affects not only people but also the species faced with potential extinction. Helen Regan and colleagues have provided an excellent overview of the types of uncertainty that arise in these contexts and suggested concrete solutions to them. They highlighted two classes of uncertainty: epistemic and linguistic. The former has to do primarily with the challenges in science of accurately and precisely measuring things that cannot be perfectly observed or counted, such as measurement error, natural variation, and potential observer bias; the latter touches on many of the issues Fischhoff highlights regarding different uses and understandings of concepts and words. I suggest this taxonomy of uncertainty as an additional resource, not as something to replace or supersede that offered by Fischhoff, because it can help to further clarify where and why problems arise for experts and decisionmakers in addressing uncertainty.

In the end, even if all aspects of uncertainty are clearly articulated and understood, there will still be potentially large differences in how people choose to operationalize that uncertainty. People have very different levels of risk tolerance, and decisionmaking under uncertainty is ultimately a proposition of taking risks. It is difficult to change people’s risk tolerance, but decisionmaking could be significantly improved if the many suggestions Fisch hoff offers were regularly implemented.

BENJAMIN S. HALPERN

Director, Center for Marine Assessment and Planning

Senior Fellow, World Conservation Monitoring Center, United Nations Environmental Programme

Research Biologist, National Center for Ecological Analysis and Synthesis

Santa Barbara, California


Baruch Fischhoff ’s excellent article on communicating uncertainty resonates with my experience in the U.S. government and prompts me to underscore two of his observations. Decisionmakers often receive a surfeit of advice but a paucity of information. Advice is usually accompanied by arguments to justify or discredit particular options. Good arguments can win debates but cannot guarantee good decisions. Good decisions, or at least well-informed decisions, require more information and better understanding of the problem. Wise decisionmakers turn to experts not because they can give better advice, but because they can enhance understanding of the matter to be decided.

Experts are most helpful—and most valuable—when they communicate what is known, what is not known, the volume and quality of information, and degrees of certainty and uncertainty. It often takes an expert to explain why abundant information sometimes provides little insight into a situation, or when experience, theory, and a few clues permit confident judgments about what is happening and why. Decisionmakers value assessments provided by experts, but they value at least as much what experts can tell them about confidence levels, probabilities, and uncertainties. If information and analysis are the ice under decisions, experts owe it to decisionmakers to tell them how thick or thin that ice is.

Experts who fail to convey uncertainties and confidence levels effectively have disserved those they are supposed to assist as surely as do experts who omit key facts or fail to challenge decisionmaker assertions they judge to be wrong. Fischhoff is correct that analysts and other experts often are reluctant to reveal what they do not know for fear that doing so will undercut their supposed expertise or convince customers that they have been less than diligent when pursuing answers to difficult questions. Conversely, decisionmakers often fail to press experts for more information about their judgments for fear of appearing less knowledgeable than they should be or risking discovering that there is little support for the position they hold.

Fischhoff also touches upon the twoaudiences problem. One audience is the decisionmaker with whom an expert works on a regular basis. Familiarity and trust facilitate clear communication, often without the need to rehearse points discussed previously. The more effectively decisionmakers communicate what they want to know, what they want to achieve, and what they think they already know, the greater the ability of experts to correct mistaken understanding, address critical issues, and ensure that the decisionmaker understands the reasons for high or low confidence in specific judgments.

The second audience consists of decisionmakers and experts who do not know one another and generally communicate indirectly. Unless precautions are taken, there is considerable risk that materials tailored for a specific and trusted customer will be misunderstood by other recipients. This is a particular problem with respect to communicating uncertainty, because what is conveyed in writing to a trusted interlocutor able to ask questions directly is likely to be too sketchy to convey uncertainties and confidence effectively to third parties. Conversely, materials prepared for general audiences risk being too diffuse to meet the needs of busy customers. This makes it even more important to provide clear confidence and probability assessments.

THOMAS FINGAR

Oksenberg-Rohlen Distinguished Fellow

Freeman Spogli Institute for International Studies

Stanford University

Stanford, California


Baruch Fischhoff and his colleagues at Carnegie Mellon University have consistently been among the strongest advocates for the proposition that acknowledging uncertainty is a strength, not a weakness, of any analysis. His article is once again in the vanguard, as he emphasizes that much of the blame for a vicious circle of overconfident assessments and precarious decisions rests at the feet of the recipients of risk and economic analyses, who have often welcomed overconfidence and rewarded candor by throwing the messengers under the proverbial bus. I especially applaud his observation that some analytic disciplines are complicit by their eagerness to “oversell their wares.” I am currently leading a National Science Foundation project on the difficulty of making sound cost/benefit decisions when regulatory economists report costs with less rigor (and especially with less attention to uncertainty) than their counterparts report risk and benefit information.

But I hope Fischhoff would agree that overconfidence is not the only trap we must avoid, and that scientific and economic information are not the only areas where uncertainty matters.

First, what Fischhoff calls “exaggerated hesitancy” he might more bluntly have called “manufactured uncertainty” (a term coined by epidemiologist David Michaels) or perhaps “strategic diffidence.” We expect some scientists with financial stakes in thwarting regulation to emphasize potentially exculpatory information (or to concoct exculpatory theories) to extend the “left-hand tail” of the uncertainty range for risk. But I believe we should also expect scientists who are public servants to resist the temptation to go along to get along by accepting more and more “underconfidence” uncritically. For example, the Environmental Protection Agency has always declined to quantify the uncertainty around its estimates of carcinogenic potency, preferring to present a single “plausible upper bound” number but to always surround it with the caveat that the potency “could be as low as zero”—whether or not that deflating verbal hedge has any reasonable basis in theory or evidence. The agency is now poised to issue a sweeping repudiation of its own 30-year track record of requiring regulated industries to present compelling evidence in order to supplant time-tested default assumptions with new assertions of safety. Sometimes, wider confidence bounds can reflect acquiescence rather than humility, and can be just as fatal to understanding as narrow or nonexistent bounds.

Second, misinterpreting expressions of uncertainty, even when they are correctly assessed, can lead decisionmakers and the public astray. Many distributions of uncertainty (and inter-individual variability) are right-skewed, but how many decisionmakers (or experts) understand that in such cases, a high percentile such as the 90th may yet be an underestimate of the expectation of the entire distribution? And most nontrivial decisions require the comparison of two or more quantities or decision alternatives: Which of two risks is larger, or which side of the decision ledger (costs versus benefits of action) is greater? Suppose the uncertain costs of reducing a particular environmental hazard follow a “normal” (bell-shaped) distribution, with a mean of $5 billion and confidence bounds of $1 billion and $9 billion. Suppose further that the monetized benefits of doing so have exactly the same uncertainty distribution (same mean and same upper and lower bounds). Is it correct to overlay and subtract the two distributions and tell the public that there is zero net benefit or net cost, regardless of whether we reduce or ignore the hazard? The fact is that unless the uncertainties happen to be tightly correlated with each other, it is much more honest to say that the consequences of controlling this hazard are so uncertain that they could benefit society by as much as $8 billion or could instead cost us as much as $8 billion (high-minus-low bounds or low-minus-high). Analysts and decisionmakers have to get better, among other examples, at explaining the vast distinction between “inconsequential” and “either terrible or terrific, with equal probability.”

I also see two areas where we need to apply Fischhoff ’s wisdom about uncertainty more broadly. When you are making a decision for someone else (or trying to provide them with the raw materials to decide), not knowing their actual preferences creates an uncertainty that is tempting to ignore or average away. There is a growing literature, for example, on the fallacy inherent in medical guidelines such as the rule of thumb that pregnant women over age 35 should strongly consider amniocentesis. This age may indeed be the point at which the miscarriage risk of the procedure itself and the age-related risk of carrying a fetus with Down syndrome are equal, but the single number is only the point where the expected harms are also equal for a woman who regards the two adverse outcomes as identically dire. There really is a distribution of tipping points that traces the distribution of personal ethical judgments across the population.

This leads to the expansion I believe Fischhoff also impatiently awaits: the marshaling of uncertainty to describe choices, not just evidence. I think we need new language that treats important decisions for what they are: attempts to minimize the probability and/or the severity of the inevitable occasions when we will choose wrongly. Only honest and comparative statements of uncertainty in outcomes can help us make the right choices for the right reasons and console us when the best we can do turns out to be not good enough. Analysts and decisionmakers need to collaborate more, not just to understand things better, but to choose more wisely. Fischhoff ’s recommen- dations put us squarely on the path to that goal.

ADAM M. FINKEL

Senior Fellow and Executive Director

Penn Program on Regulation

University of Pennsylvania Law School

Philadelphia, Pennsylvania


The trouble with STEM

It has become popular to combine science, technology, engineering, and mathematics into the acronym STEM when discussing education (see Robert D. Atkinson’s article “Why the Current Education Reform Strategy Won’t Work,” Issues, Spring 2012). But as an engineer, I have some difficulty with STEM. Science requires curiosity to produce new knowledge, technology requires skill, mathematics requires logic, but engineering requires creativity and judgment. Engineering, to be sure, requires science, technology, and mathematics (STM) as tools to ensure that a design will be successful, but the first step in the design process is invention: the critical beginning of a new idea. In this respect, engineering is different from STM.

As a child, I had many interests and dreamed that someday I would become either a scientist or an artist. That dream persisted until high school, when I learned about engineering. The path I chose combined both of my passions into one profession.

The humanitarian and artistic side of engineering is often downplayed, if not ignored, in engineering education. Engineering science courses, essentially applied physics, use mathematical tools almost exclusively to solve problems. These engineering science courses give nascent engineers the technological skills to analyze a trial design to assess the likelihood of success. But the design process involves synthesis as well as analysis, and the synthesis part of this process is not given the same attention as the analysis part. We engineering educators depend very much on the native creativity of our students to be able to synthesize.

Along with synthesis, engineers must learn judgment: how to go about the process of determining whether or not a new concept will work, often before the mathematical analysis stage. Judgment requires experience, and experience usually involves failure on somebody’s part. If that failure does not happen in school, where its consequences are limited, then that failure and its accompanying experience and growth of engineering judgment has to be developed in the workplace. That is why employers of just-graduated engineers often give their new employees mentored dummy projects to work on during the first six months of their tenure. An engineer who has not developed engineering judgment by that time is either not retained or faces potentially more catastrophic failure later in her/his career.

The Accreditation Board for Engineering and Technology requires the inclusion of humanities courses in engineering curricula. The reason is clear: Engineers often need to include nonscientific, non-mathematical, even nontechnological features in their designs. There are artistic, historical, political, psychological, financial, spatial, and other elements that must be considered for a prototype engineering design to be successful. There have been many examples of new products that have failed because they didn’t look right, were named wrong, didn’t feel right, or were too complicated. All of these fit into the category of engineering judgment.

The reason that the “E” in STEM should be distinguished from the STM part is that engineering can appeal to those who are not so good at STM, but who have superior skills or interest in the softer side of engineering. They can become great patent lawyers, engineering managers, recruiters, teachers, insurance adjusters, human factors engineers, or sales/applications engineers. Some engineering disciplines are more people-oriented than others. I have had “C” students in my Transport Processes and Electronic Design classes who could talk a lawyer into the ground. With the technical backgrounds that they had barely mastered, they used the softer skills in which they excelled to find successful careers on the fringes of engineering. We need these kinds of people, too.

The American Association of University Women’s 2010 report Why So Few? Women in Science, Technology, Engineering, and Mathematics deplores the fact that it is difficult to attract young women into STEM fields. One reason they give for this is the stronger interest young women have in people and interpersonal relationships that are not usually associated with STEM fields. But if it were better known that engineers do not have to be nerds and recluses, that as engineers they could have a rich involvement with other people, then the “E” part of STEM might attract more women and men to the profession. This is the problem with the acronym STEM; it does not do justice to the full range of opportunities available in engineering.

ARTHUR T. JOHNSON

Professor Emeritus

Fischell Department of Bioengineering

University of Maryland

College Park, Maryland

Applying New Research to Improve Science Education

Science, technology, engineering, and mathematics (STEM) education is critical to the U.S. future because of its relevance to the economy and the need for a citizenry able to make wise decisions on issues faced by modern society. Calls for improvement have become increasingly widespread and desperate, and there have been countless national, local, and private programs aimed at improving STEM education, but there continues to be little discernible change in either student achievement or student interest in STEM. Articles and letters in the spring and summer 2012 editions of Issues extensively discussed STEM education issues. Largely absent from these discussions, however, is attention to learning.

This is unfortunate because there is an extensive body of recent research on how learning is accomplished, with clear implications for what constitutes effective STEM teaching and how that differs from typical current teaching at the K-12 and college levels. Failure to understand this learning-focused perspective is also a root cause of the failures of many reform efforts. Furthermore, the incentive systems in higher education, in part driven by government programs, act to prevent the adoption of these research-based ideas in teaching and teacher training.

A new approach

The current approach to STEM education is built on the assumption that students come to school with different brains and that education is the process of immersing these brains in knowledge, facts, and procedures, which those brains then absorb to varying degrees. The extent of absorption is largely determined by the inherent talent and interest of the brain. Thus, those with STEM “talent” will succeed, usually easily, whereas the others have no hope. Research advances in cognitive psychology, brain physiology, and classroom practices are painting a very different picture of how learning works.

We are learning that complex expertise is a matter not of filling up an existing brain with knowledge, but of brain development. This development comes about as the result of intensive practice of the cognitive processes that define the specific expertise, and effective teaching can greatly reduce the impact of initial differences among the learners.

This research has established important underlying causes and principles and important specific results, but it is far from complete. More research is needed on how to accomplish the desired learning most effectively over the full range of STEM skills and potential learners in our classrooms, as well as how to best train teachers.

What is learning STEM?

The appropriate STEM educational goal should be to maximize the extent to which the learners develop expertise in the relevant subject, where expertise is defined by what scientists and engineers do. This is not to say that every learner should become a scientist or engineer, or that they could become one by taking any one class, but rather that the value of the educational experiences should be measured by their effectiveness at changing the thinking of the learner to be more like that of an expert when solving problems and making decisions relevant to the discipline. As discussed in the National Research Council study Taking Science to School, modern research has shown that children have the capability to begin this process and learn complex reasoning at much earlier ages than previously thought, at least from the beginning of their formal schooling. Naturally, it is necessary and desirable for younger children to learn less specialized expertise encompassing a broader range of disciplines than would be the case for older learners.

Expertise has been extensively studied across a variety of disciplines. Experts in any given discipline have large amounts of knowledge and particular discipline-specific ways in which they organize and apply that knowledge. Experts also have the capability to monitor their own thinking when solving problems in their discipline, testing their understanding and the suitability of different solution approaches, and making corrections as appropriate. There are a number of more specific components of expertise that apply across the STEM disciplines. These include the use of:

  • Discipline- and topic-specific mental models involving relevant cause and effect relationships that are used to make predictions about behavior and solve problems.
  • Sophisticated criteria for deciding which of these models do or don’t apply in a given situation, and processes for regularly testing the appropriateness of the model being used.
  • Complex pattern-recognition systems for distinguishing between relevant and irrelevant information.
  • Specialized representations.
  • Criteria for selecting the likely optimum solution method to a given problem.
  • Self-checking and sense making, including the use of discipline-specific criteria for checking the suitability of a solution method and a result.
  • Procedures and knowledge, some discipline-specific and some not, that have become so automatic with practice that they can be used without requiring conscious mental processing. This frees up cognitive resources for other tasks.

Many of these components involve making decisions in the presence of limited information—a vital but often educationally neglected aspect of expertise. All of these components are embedded in the knowledge and practices of the discipline, but that knowledge is linked with the process and context, which are essential elements for knowledge to be useful. Similarly, measuring the learning of most elements of this expertise is inherently discipline-specific.

How is learning achieved?

Researchers are also making great progress in determining how expertise is acquired, with the basic conclusion being that those cognitive processes that are explicitly and strenuously practiced are those that are learned. The learning of complex expertise is thus quite analogous to muscle development. In response to the extended strenuous use of a muscle, it grows and strengthens. In a similar way, the brain changes and develops in response to its strenuous extended use. Advances in brain science have now made it possible to observe some of these changes.

Specific elements, collectively called “deliberate practice,” have been identified as key to acquiring expertise across many different areas of human endeavor. This involves the learner solving a set of tasks or problems that are challenging but doable and that involve explicitly practicing the appropriate expert thinking and performance. The tasks must be sufficiently difficult to require intense effort by the learner if progress is to be made, and hence must be adjusted to the current state of expertise of the learner. Deliberate practice also includes internal reflection by the learner and feedback from the teacher/coach, during which the achievement of the learner is compared with a standard, and there is an analysis of how to make further progress. The level of expert-like performance has been shown to be closely linked to the duration of deliberate practice. Thousands of hours of deliberate practice are typically required to reach an elite level of performance.

This research has a number of important implications for STEM education. First, it means that learning is inherently difficult, so that motivation plays a large role. To succeed, the learner must be convinced of the value of the goal and believe that hard work, not innate talent, is critical. Second, activities that do not demand substantial focus and effort provide little educational value. Listening passively to a lecture, doing many easy, repetitive tasks, or practicing irrelevant skills produce little learning. Third, although there are distinct differences among learners, for the great majority the amount of time spent in deliberate practice transcends any other variables in determining learning outcomes.

Implications for teaching

From the learning perspective, effective teaching is that which maximizes the learner’s engagement in cognitive processes that are necessary to develop expertise. As such, the characteristics of an effective teacher are very analogous to those of a good athletic coach: designing effective practice activities that break down and collectively embody all the essential component skills, motivating the learner to work hard on them, and providing effective feedback.

The effective STEM teacher must:

  • Understand expert thinking and design suitable practice tasks.
  • Target student thinking and learning needs. Such tasks must be appropriate to the level of the learner and be effective at building on learners’ current thinking to move them to higher expertise. The teacher must be aware of and connect with the prior thinking of the learner as well as have an understanding of the cognitive difficulties posed by the material.
  • Motivate the student to put in the extensive effort that is required for learning. This involves generating a sense of self-efficacy and ownership of the learning; making the subject interesting, relevant, and inspiring; developing a sense of identity in the learner as a STEM expert; and other factors that affect motivation. How to do this in practice is dependent on the subject matter and the characteristics of the learner—their prior experience, level of mastery, and individual and sociocultural values.
  • Provide effective feedback that is timely and directly addresses the student’s thinking. This requires the teacher to recognize the student’s thought processes, be aware of the typical cognitive challenges with the material, and prepare particular questions, tasks, and examples to help the learner overcome those challenges. Research has shown several effective means of providing feedback, including short, focused lectures if the student has been carefully prepared to learn from that lecture.
  • Understand how learning works, and use that to guide all of their activities. In addition to the research on learning expertise, this includes other well-established principles regarding how the human brain processes and remembers information that are relevant to education, such as the limitations of the brain’s short-term memory and what processes enhance long-term retention.

Although many of these instructional activities are easier to do one on one, there are a variety of pedagogical techniques and simple technologies that extend the capabilities of the teacher to provide these elements of instruction to many students at once in a classroom, often by productively using student-student interactions. Examples of approaches that have demonstrated their effectiveness can be found in recommended reading articles by Michelle Smith and by Louis Deslauriers et al.

Effective STEM teaching is a specific learned expertise that includes, and goes well beyond, STEM subject expertise. Developing such teaching expertise should be the focus of STEM teacher training. Teachers must have a deep mastery of the content so they know what expert thinking is, but they also must have “pedagogical content knowledge.” This is an understanding of how students learn the particular content and the challenges and opportunities for facilitation of learning at a topic-specific level.

This view of STEM teaching as optimizing the development of expertise provides clearer and more detailed guidance than what is currently available from the classroom research on effective teaching. Most of the classroom research on effective teaching looks at K-12 classrooms and attempts to link student progress on standardized tests with various teacher credentials, traits, or training. Although there has been progress, it is limited because of the challenges of carrying out educational research of this type. There are a large number of uncontrolled variables in the K-12 school environment that affect student learning, the standardized tests are often of questionable validity for measuring learning, teacher credentials and training are at best tenuous measures of their content mastery and pedagogical content mastery, and the general level of these masteries is low in the K-12 teacher population. The level of mastery is particularly low in elementary- and middle-school teachers. All of these factors conspire to make the signals small and easily masked by other variables.

At the college level, the number of uncontrolled variables is much smaller, and as reviewed in the NRC report Discipline-Based Education Research, it is much clearer that those teachers who practice pedagogy that supports deliberate practice by the students show substantially greater learning gains than are achieved with traditional lectures. For example, the learning of concepts for all students is improved, with typical increases of 50 to 100%, and the dropout and failure rates are roughly halved.

Shortcomings of the current system

Typical K-16 STEM teaching contrasts starkly with what I have just described as effective teaching. At the K-12 level, although there are notable exceptions, the typical teacher starts out with a very weak idea of what it means to think like a scientist or engineer. Very few K-12 teachers, including many who were STEM majors, acquire sufficient domain expertise in their preparation. Hence, the typical teacher begins with very little capability to properly design the requisite learning tasks. Furthermore, their lack of content mastery, combined with a lack of pedagogical content knowledge, prevents them from properly evaluating and guiding the students’ thinking. Much of the time, students in class are listening passively or practicing procedures that neither have the desired cognitive elements nor require the level of strenuousness that are important for learning.

Teachers at both the K-12 and undergraduate levels also have limited knowledge of the learning process and what is known about how the mind functions, resulting in common educational practices that are clearly counter to what research shows is optimum, both for processing and learning information in the classroom environment and for achieving long-term retention. Another shortcoming of teaching at all levels is the strong tendency to teach “anti-creativity.” Students are taught and tested on solving well-defined artificial problems posed by the teacher, where the goal is to use the specific procedure the teacher intended to produce the intended answer. This requires essentially the opposite cognitive process from STEM creativity, which is primarily recognizing the relevance of previously unappreciated relationships or information to solve a problem in a novel way.

At the undergraduate level, STEM teachers generally have a high degree of subject expertise. Unfortunately, this is not reflected in the cognitive activities of the students in the classroom, which again consist largely of listening, with very little cognitive processing needed or possible. Students do homework and exam problems that primarily involve practicing solution procedures, albeit complex and/or mathematically sophisticated ones. However, the assigned problems almost never explicitly require the sorts of cognitive tasks that are the critical components of expertise described above. Instructors also often suffer from “expert blindness,” failing to recognize and make explicit many mental processes that they have practiced so much that they are automatic.

Another problem at the postsecondary level is the common belief that effective teaching is only a matter of providing information to the learner, with everything else being the responsibility of the learners and/or their innate limitations. It is common to assume that motivation, and even curiosity about a subject, are entirely the responsibility of the student, even when the student does not yet know much about the subject.

Failure of reform efforts

The perspective on learning that I have described also explains the failure of many STEM reform efforts.

Belief in the importance of innate talent or other characteristics. Schools have long focused educational resources on learners that have been identified in some manner as exceptional. Although the research shows that all brains learn expertise in fundamentally the same way, that is not to say that all learners are the same. Many different aspects affect the learning of a particular student. Previous learning experiences and sociocultural background and values obviously play a role. There is a large and contentious literature as to the relative significance of innate ability/talent or the optimum learning style of each individual, with many claims and fads supported by little or questionable research.

Researchers have tried for decades to demonstrate that success is largely determined by some innate traits and that by measuring those traits with IQ tests or other means, one can preselect children who are destined for greatness and then focus educational resources on them. This field of research has been plagued by difficulties with selection bias and the lack of adequate controls. Although there continues to be some debate, the bulk of the research is now showing that, excepting the lower tail of the distribution consisting of students with pathologies, the predictive value of any such early tests of intellectual capability is very limited. From an educational policy point of view, the most important research result is that any predictive value is small compared to the later effects of the amount and quality of deliberate practice undertaken by the learner. That predictive value is also small compared to the effects of the learners’ and teachers’ beliefs about learning and the learners’ intellectual capabilities. Although early measurements of talent, or IQ, independent of other factors have at best small correlation with later accomplishment, simply labeling someone as talented or not has a much larger correlation. It should be noted that in many schools students who are classified as deficient by tests with very weak predictive value are put into classrooms that provide much less deliberate practice than the norm, whereas the opposite is true for students who are classified as gifted. The subsequent difference in learning outcomes for the two groups provides an apparent validation for what is merely a self-fulfilling prophecy. Given these findings, human capital is clearly maximized by assuming that, except for students with obvious pathologies, every student is capable of great achievement in STEM and should be provided with the educational experiences that will maximize their learning.

The idea that for each individual there is a unique learning style is surprisingly widespread given the lack of supporting evidence for this claim, and in fact significant evidence showing the contrary, as reviewed by Hal Pashler of the University of California at San Diego and others.

Because of the presence of many different factors that influence a student’s success in STEM, including the mind’s natural tendency to learn, some students do succeed in spite of the many deficiencies in the educational system. Most notably, parents can play a major role in both early cognitive development and STEM interest, which are major contributors to later success. However, optimizing the teaching as I described would allow success for a much larger fraction of the population, as well as allowing those students who are successful in the current system to do even better.

Poor standards and accountability. Standards have had a major role in education reform efforts, but they are very much a double-edged sword. Although good definitions and assessments of the desired learning are essential, bad definitions are very harmful. There are tremendous pitfalls in developing good, widely used standards and assessments. The old concept of learning, combined with expert blindness and individual biases, exerts a constant pressure on standards to devolve into a list of facts covering everyone’s areas of interest, with little connection to the essential elements of expertise. The shortcomings in the standards are then reinforced by the large-scale assessment systems, because measuring a student’s knowledge of memorized facts and simple procedures is much cheaper and easier than authentic measurements of expertise. So although good standards and good assessment must be at the core of any serious STEM education improvement effort, poor standards and poor assessments can have very negative consequences. The recent National Academy of Sciences–led effort on new science standards, starting with a carefully thought-out guiding framework, is an excellent start, but this must avoid all the pitfalls as it is carried through to large-scale assessments of student mastery. Finally, good standards and assessments will never by themselves result in substantial improvement in STEM education, because they are only one of several essential components to achieving learning.

Competitions and other informal science programs: Attempting to separate the inspiration from the learning. Motivation in its entirety, including the elements of inspiration, is such fundamental requirement for learning that any approach that separates it from any aspect the learning process is doomed to be ineffective. Unfortunately, a large number of government and private programs that support the many science and engineering competitions and out-of-school programs assume that they are separable. The assumption of such programs is that by inspiring children through competitions or other enrichment experiences, they will then thrive in formal school experiences that provide little motivation or inspiration and still go on to achieve STEM success. Given the questionable assumptions about the learning process that underlie these programs, we should not be surprised that there is little evidence that such programs ultimately succeed, and some limited evidence to the contrary. The past 20 years have seen an explosion in the number of participants in engineering-oriented competitions such as First Robotics and others, while the fraction of the population getting college degrees in engineering has remained constant. A study by Rena Subotnik and colleagues that tracked high-school Westinghouse (now Intel) talent search winners, an extraordinarily elite group already deeply immersed in science, found that a substantial fraction, including nearly half of the women, had switched out of science within a few years, largely because of their experiences in the formal education system. It is not that such enrichment experiences are bad, just that they are inherently limited in their effectiveness. Programs that introduce these motivational elements as an integral part of every aspect of the STEM learning process, particularly in formal schooling, would probably be more effective.

Silver-bullet solutions. A number of prominent scientists, beginning as far back as the Sputnik era, have introduced new curricula based on their understanding of the subject. The implicit assumption of such efforts is that someone with a high level of subject expertise can simply explain to novices how an expert thinks about the subject, and the novices (either students or K-12 teachers) will then embrace and use that way of thinking and be experts themselves. This assumption is strongly contradicted by the research on expertise and learning, and so the failure of such efforts is no surprise.

A number of elements such as school organization, teacher salaries, working conditions, and others have been put forth as the element that, if changed, will fix STEM education. Although some of these may well be a piece of a comprehensive reform, they are not particularly STEM-specific and by themselves will do little to address the basic shortcomings in STEM teaching and learning.

The conceptual flaws of STEM teacher in-service professional development. The federal government spends a few hundred million dollars each year on in-service teacher professional development in STEM, with states and private sources providing additional funding. Suzanne Wilson’s review of the effectiveness of such professional development activities finds evidence of little success and identifies structural factors that inhibit effectiveness. From the perspective of learning expertise, it is clear why teacher professional development is fundamentally ineffective and expensive. If these teachers failed to master the STEM content as full-time students in high school and college, it is unrealistic to think they will now achieve that mastery as employees through some intermittent, part-time, usually voluntary activity on top of their primary job.

Why change is hard

First, nearly everyone who has gone to school perceives himself or herself to be an expert on education, resulting in a tendency to seize on solutions that overlook the complexities of the education system and how the brain learns. Second, there are long-neglected structural elements and incentives within the higher education system that actively inhibit the adoption of better teaching methods and the better training of teachers. These deserve special attention.

Improving undergraduate STEM teaching to produce better-educated graduates and better-trained future K-12 teachers is a necessary first step in any serious effort to improve STEM education, but there are several barriers to accomplishing this. First, the tens of billions of dollars of federal research funding going to academic institutions, combined with no accountability for educational outcomes at the levels of the department or the individual faculty member, have shaped the university incentive system to focus almost entirely on research. Thus, STEM departments and individual faculty members, regardless of their personal inclinations, are forced to prioritize their time accordingly, with the adoption of better teaching practices, improved student outcomes, and contributing to the training of future K-12 STEM teachers ranking very low. Second, to the limited extent that there are data, STEM instructional practices appear to be similarly poor across the range of postsecondary institutions. This is probably because the research-intensive universities produce most of the Ph.D.s, who become the faculty at all types of institutions, and so the educational values and standards of the research-intensive universities have become pervasive. Third, with a few exceptions, the individual academic departments retain nearly absolute control over what they teach and how they teach. Deans, provosts, and especially presidents have almost no authority over, or even knowledge of, educational practices in use by the faculty. Any successful effort to change undergraduate STEM teaching must change the incentives and accountability at the level of the academic department and the individual faculty member in the research-intensive universities.

A possible option would be to make a department’s eligibility to receive federal STEM research funds contingent on the reporting and publication of undergraduate teaching practices and student outcomes. A standard reporting format would make it possible to compare the extent to which departments and institutions employ best practices. Prospective students could then make more-informed decisions about which institution and department would provide them with the best education.

Most K-12 teacher preparation programs have a local focus, and they make money for the institutions of which they are a part. There is no accepted professional standard for teacher training, and there is a financial incentive for institutions to accept and graduate as many education majors as possible. This has resulted in low standards, particularly in math and science, with teacher education programs frequently having the lowest math and science requirements of any major at the institution. This also means that they attract students with the greatest antipathy toward math and science. Research by my colleagues has found that elementary education majors have far more novice-like attitudes about physics than do students in any other major at the university. Federal programs to support the training of K-12 STEM teachers provide easily available scholarship money, which reinforces the status quo by ensuring a plentiful supply of students in spite of the programs’ low quality. Rewarding institutions that produce graduates with the expertise needed to be highly effective teachers is an essential step in bringing about the massive change that is needed in the preparation of STEM teachers.

Focusing on STEM learning and how it is achieved provides a valuable perspective for understanding the shortcomings of the educational system and how it can be improved. It clarifies why the current system is producing poor results and explains why current and past efforts to improve the situation have had little effect. However, it also offers hope. Improvement is contingent on changes in the incentive system in higher education to bring about the widespread adoption of STEM teaching methods and the training of K-12 teachers that embody what research has shown is important for effective learning. These tasks are admittedly challenging, but the results would be dramatic. The United States would go from being a laggard in STEM education to the world leader.