b Understanding Iran beyond the deal By webfeeds.brookings.edu Published On :: Mon, 30 Nov -0001 00:00:00 +0000 On October 15, the Center for Middle East Policy hosted a conversation with Suzanne Maloney, deputy director of Brookings Foreign Policy program and author of the recently released book, Iran’s Political Economy since the Revolution (Cambridge University Press, 2015); Javier Solana, Brookings distinguished fellow and former EU High Representative for the Common Foreign and Security Policy; and Vali Nasr, Dean of Johns Hopkins University School of Advanced International Studies and nonresident senior fellow at Brookings. The three experts discussed Iran today, the implications of the nuclear agreement, and more. Full Article
b In the Wake of BCRA: An Early Report on Campaign Finance in the 2004 Elections By webfeeds.brookings.edu Published On :: Tue, 15 Jun 2004 00:00:00 -0400 ABSTRACT: Early experience with federal campaign finance reform suggests that the new law is fulfilling its primary objective of severing links between policymakers and large donors, and thus reducing the potential for corruption in the political process. Instead of languishing or seeking to circumvent the law, the national political parties have responded to the ban on soft money by increasing their hard money resources. While outside groups appear active, particularly on the Democratic side, their soft money financing should remain a small fraction of what candidates and parties will raise and spend in the 2004 Elections.To read the full article, please visit The Forum's website Authors Anthony CorradoThomas E. Mann Publication: The Forum Full Article
b Despite Predictions, BCRA Has Not Been a Democratic 'Suicide Bill' By webfeeds.brookings.edu Published On :: Mon, 26 Jul 2004 00:00:00 -0400 During debates in Congress and in the legal battles testing its constitutionality, critics of the Bipartisan Campaign Reform Act of 2002 imagined a host of unanticipated and debilitating consequences. The law's ban on party soft money and the regulation of electioneering advertising would, they warned, produce a parade of horribles: A decline in political speech protected by the First Amendment, the demise of political parties, and the dominance of interest groups in federal election campaigns.The forecast that attracted the most believers — among politicians, journalists, political consultants, election-law attorneys and scholars — was the claim that Democrats would be unable to compete against Republicans under the new rules, primarily because the Democrats' relative ability to raise funds would be severely crippled. One year ago, Seth Gitell in The Atlantic Monthly summarized this view and went so far as to call the new law "The Democratic Party Suicide Bill." Gitell quoted a leading Democratic Party attorney, who expressed his private view of the law as "a fascist monstrosity." He continued, "It is grossly offensive ... and on a fundamental level it's horrible public policy, because it emasculates the parties to the benefit of narrow-focus special-interest groups. And it's a disaster for the Democrats. Other than that, it's great."The core argument was straightforward. Democratic Party committees were more dependent on soft money — unlimited contributions from corporations, unions and individuals — than were the Republicans. While they managed to match Republicans in soft-money contributions, they trailed badly in federally limited hard-money contributions. Hence, the abolition of soft money would put the Democrats at a severe disadvantage in presidential and Congressional elections.In addition, the argument went, by increasing the amount an individual could give to a candidate from $1,000 to $2,000, the law would provide a big financial boost to President Bush, who would double the $100 million he raised in 2000 and vastly outspend his Democratic challenger. Finally, the ban on soft money would weaken the Democratic Party's get-out-the-vote efforts, particularly in minority communities, while the regulation of "issue ads" would remove a potent electoral weapon from the arsenal of labor unions, the party's most critical supporter.After 18 months of experience under the law, the fundraising patterns in this year's election suggest that these concerns were greatly exaggerated. Money is flowing freely in the campaign, and many voices are being heard. The political parties have adapted well to an all-hard-money world and have suffered no decline in total revenues. And interest groups are playing a secondary role to that of the candidates and parties.The financial position of the Democratic party is strikingly improved from what was imagined a year ago. Sen. John Kerry (D-Mass.), who opted out of public funding before the Iowa caucuses, will raise more than $200 million before he accepts his party's nomination in Boston. The unusual unity and energy in Democrats' ranks have fueled an extraordinary flood of small donations to the Kerry campaign, mainly over the Internet. These have been complemented by a series of successful events courting $1,000 and $2,000 donors.Indeed, since Kerry emerged as the prospective nominee in March, he has raised more than twice as much as Bush and has matched the Bush campaign's unprecedented media buys in battleground states, while also profiting from tens of millions of dollars in broadcast ads run by independent groups that are operating largely outside the strictures of federal election law.The Democratic national party committees have adjusted to the ban on soft money much more successfully than insiders had thought possible. Instead of relying on large soft-money gifts for half of their funding, Democrats have shown a renewed commitment to small donors and have relied on grassroots supporters to fill their campaign coffers. After the 2000 election, the Democratic National Committee had 400,000 direct-mail donors; today the committee has more than 1.5 million, and hundreds of thousands more who contribute over the Internet.By the end of June, the three Democratic committees had already raised $230 million in hard money alone, compared to $227 million in hard and soft money combined at this point in the 2000 election cycle. They have demonstrated their ability to replace the soft money they received in previous elections with new contributions from individual donors.Democrats are also showing financial momentum as the election nears, and thus have been gradually reducing the Republican financial advantage in both receipts and cash on hand. In 2003, Democrats trailed Republicans by a large margin, raising only $95 million, compared to $206 million for the GOP. But in the first quarter of this year, Democrats began to close the gap, raising $50 million, compared to $82 million for Republicans. In the most recent quarter, they narrowed the gap even further, raising $85 million, compared to the Republicans' $96 million.Democrats are now certain to have ample funds for the fall campaigns. Although they had less than $20 million in the bank (minus debts) at the beginning of this year, they have now banked $92 million. In the past three months, Democrats actually beat Republicans in generating cash — $47 million, compared to $31 million for the GOP.The party, therefore, has the means to finance a strong coordinated and/or independent-spending campaign on behalf of the presidential ticket, while Congressional committees have the resources they need to play in every competitive Senate and House race, thanks in part to the fundraising support they have received from Members of Congress.Moreover, FEC reports through June confirm that Democratic candidates in those competitive Senate and House races are more than holding their own in fundraising. They will be aided by a number of Democratic-leaning groups that have committed substantial resources to identify and turn out Democratic voters on Election Day.Democrats are highly motivated to defeat Bush and regain control of one or both houses of Congress. BCRA has not frustrated these efforts. Democrats are financially competitive with Republicans, which means the outcome will not be determined by a disparity of resources. Put simply, the doomsday scenario conjured up by critics of the new campaign finance law has not come to pass. Authors Anthony CorradoThomas E. Mann Publication: Roll Call Full Article
b The New Campaign Finance Sourcebook By webfeeds.brookings.edu Published On :: Sat, 01 Oct 2005 00:00:00 -0400 Brookings Institution Press 2005 292pp. This completely revised and expanded update of Campaign Finance Reform: A Sourcebook provides the definitive exposition of federal campaign finance regulation. Written by four of the nation's most influential analysts on politics and money, The New Campaign Finance Sourcebook presents a thorough overview and analysis of campaign finance policy and practices, including the history of campaign finance regulation state of campaign finance law and the implementation of BCRA constitutional and regulatory issues in the campaign finance debate current practices and trends in the financing of federal elections public financing of presidential elections rules for campaigning on the internet alternative approaches to reform. The New Campaign Finance Sourcebook has also been integrated with the popular and useful Brookings website on campaign finance to provide a timely, interactive tool for policymakers, journalists, campaign professionals, and scholars. The Brookings Institution has been a leader in analyzing campaign finance and this important new book is an essential addition to that proud tradition. ABOUT THE AUTHORS Anthony Corrado Daniel R. Ortiz Daniel R. Ortiz is the John Allan Love Professor of Law and Horace W. Goldsmith Research Professor at the University of Virginia School of Law. Thomas E. Mann Trevor Potter Downloads Sample Chapter Ordering Information: {9ABF977A-E4A6-41C8-B030-0FD655E07DBF}, 978-0-8157-0005-0, $26.95 Add to Cart Full Article
b @Brookings Podcast: The Influence of Super PACs on the 2012 Elections By webfeeds.brookings.edu Published On :: Fri, 09 Mar 2012 16:20:00 -0500 Super PACs have already spent tens of millions of dollars in the race for the GOP presidential nomination, with more to come. Expert Anthony Corrado says that the unlimited spending by the PACs, made possible by two Supreme Court decisions, is giving wealthy individuals unprecedented influence in the 2012 elections. previous play pause next mute unmute @Brookings Podcast: The Influence of Super PACs on the 2012 Elections 07:13 Download (Help) Get Code Brookings Right-click (ctl+click for Mac) on 'Download' and select 'save link as..' Get Code Copy and paste the embed code above to your website or blog. Video The Influence of Super PACs on the 2012 Elections Audio @Brookings Podcast: The Influence of Super PACs on the 2012 Elections Image Source: © Jessica Rinaldi / Reuters Full Article
b Beyond great forces: How individuals still shape history By webfeeds.brookings.edu Published On :: Tue, 15 Oct 2019 19:09:44 +0000 Full Article
b Webinar: Telehealth before and after COVID-19 By webfeeds.brookings.edu Published On :: Mon, 27 Apr 2020 14:35:44 +0000 The coronavirus outbreak has generated an immediate need for telehealth services to prevent further infections in the delivery of health care. Before the global pandemic, federal and state regulations around reimbursement and licensure requirements limited the use of telehealth. Private insurance programs and Medicaid have historically excluded telehealth from their coverage, and state parity laws… Full Article
b COVID-19 has taught us the internet is critical and needs public interest oversight By webfeeds.brookings.edu Published On :: Wed, 29 Apr 2020 17:50:42 +0000 The COVID-19 pandemic has graphically illustrated the importance of digital networks and service platforms. Imagine the shelter-in-place reality we would have experienced at the beginning of the 21st century, only two decades ago: a slow internet and (because of that) nothing like Zoom or Netflix. Digital networks that deliver the internet to our homes, and… Full Article
b COVID-19 trends from Germany show different impacts by gender and age By webfeeds.brookings.edu Published On :: Fri, 01 May 2020 15:41:03 +0000 The world is in the midst of a global pandemic and all countries have been impacted significantly. In Europe, the most successful policy response to the pandemic has been by Germany, as measured by the decline in new COVID-19 cases in recent weeks and consistent increase in recovered’ cases. This is also reflected in the… Full Article
b Removing regulatory barriers to telehealth before and after COVID-19 By webfeeds.brookings.edu Published On :: Wed, 06 May 2020 16:00:55 +0000 Introduction A combination of escalating costs, an aging population, and rising chronic health-care conditions that account for 75% of the nation’s health-care costs paint a bleak picture of the current state of American health care.1 In 2018, national health expenditures grew to $3.6 trillion and accounted for 17.7% of GDP.2 Under current laws, national health… Full Article
b How to increase financial support during COVID-19 by investing in worker training By webfeeds.brookings.edu Published On :: Wed, 06 May 2020 17:46:07 +0000 It took just two weeks to exhaust one of the largest bailout packages in American history. Even the most generous financial support has limits in a recession. However, I am optimistic that a pandemic-fueled recession and mass underemployment could be an important opportunity to upskill the American workforce through loans for vocational training. Financially supporting… Full Article
b Obama’s legacy in African security and development By webfeeds.brookings.edu Published On :: Mon, 25 Jul 2016 16:36:00 +0000 President Obama’s presidency has witnessed widespread change throughout Africa. What legacy will he leave on the continent? Full Article Uncategorized
b Was Saudi King Salman too sick to attend this week’s Arab League summit? By webfeeds.brookings.edu Published On :: Mon, 30 Nov -0001 00:00:00 +0000 King Salman failed to show at the Arab League summit this week in Mauritania, allegedly for health reasons. The king’s health has been a question since his accession to the throne last year. Full Article Uncategorized
b Hey, Kremlin: Americans can make loose talk about nukes, too By webfeeds.brookings.edu Published On :: Thu, 04 Aug 2016 16:29:21 +0000 Over the past several years, Vladimir Putin and senior Russian officials have talked loosely about nuclear weapons, suggesting the Kremlin might not fully comprehend the awful consequences of their use. That has caused a degree of worry in the West. Now, the West has in Donald Trump—the Republican nominee to become the next president of […] Full Article
b The Marketplace of Democracy: A Groundbreaking Survey Explores Voter Attitudes About Electoral Competition and American Politics By webfeeds.brookings.edu Published On :: Fri, 27 Oct 2006 10:00:00 -0400 Event Information October 27, 200610:00 AM - 12:00 PM EDTFalk AuditoriumThe Brookings Institution1775 Massachusetts Ave., NWWashington, DC Register for the EventDespite the attention on the mid-term races, few elections are competitive. Electoral competition, already low at the national level, is in decline in state and primary elections as well. Reformers, who point to gerrymandering and a host of other targets for change, argue that improving competition will produce voters who are more interested in elections, better-informed on issues, and more likely to turn out to the polls. On October 27, the Brookings Institution—in conjunction with the Cato Institute and The Pew Research Center—presented a discussion and a groundbreaking survey exploring the attitudes and opinions of voters in competitive and noncompetitive congressional districts. The survey, part of Pew's regular polling on voter attitudes, was conducted through the weekend of October 21. A series of questions explored the public's perceptions, knowledge, and opinions about electoral competitiveness. The discussion also explored a publication that addresses the startling lack of competition in our democratic system. The Marketplace of Democracy: Electoral Competition and American Politics (Brookings, 2006), considers the historical development, legal background, and political aspects of a system that is supposed to be responsive and accountable, yet for many is becoming stagnant, self-perpetuating, and tone-deaf. Michael McDonald, editor and Brookings visiting fellow, moderated a discussion among co-editor John Samples, director of the Center for Representative Government at the Cato Institute, and Andrew Kohut and Scott Keeter from The Pew Research Center, who also discussed the survey. Transcript Transcript (.pdf) Event Materials 2006102720061027ppt Full Article
b Five Myths About Turning Out the Vote By webfeeds.brookings.edu Published On :: Sun, 29 Oct 2006 00:00:00 -0400 If you're an upstanding U.S. citizen, you'll stand up and be counted this Election Day, right? Well, maybe not. Just because Americans can vote doesn't mean they do. But who shows up is what decides the tight races, which makes turnout one of the most closely watched aspects of every election -- and one that has fostered a number of myths. Here are five, debunked:1. Thanks to increasing voter apathy, turnout keeps dwindling. This is the mother of all turnout myths. There may be plenty of apathetic voters out there, but the idea that ever fewer Americans are showing up at the polls should be put to rest. What's really happening is that the number of people not eligible to vote is rising -- making it seem as though turnout is dropping. Those who bemoan a decline in American civic society point to the drop in turnout from 55.2 percent in 1972, when 18-year-olds were granted the right to vote, to the low point of 48.9 percent in 1996. But that's looking at the total voting-age population, which includes lots of people who aren't eligible to vote -- namely, noncitizens and convicted felons. These ineligible populations have increased dramatically over the past three decades, from about 2 percent of the voting-age population in 1972 to 10 percent today. When you take them out of the equation, the post-1972 "decline" vanishes. Turnout rates among those eligible to vote have averaged 55.3 percent in presidential elections and 39.4 percent in midterm elections for the past three decades. There has been variation, of course, with turnout as low as 51.7 percent in 1996 and rebounding to 60.3 percent by 2004. Turnout in the most recent election, in fact, is on a par with the low-60 percent turnout rates of the 1950s and '60s. 2. Other countries' higher turnout indicates more vibrant democracies. You can't compare apples and oranges. Voting rules differ from nation to nation, producing different turnout rates. Some countries have mandatory voting. If Americans were fined $100 for playing voter hooky on Election Day, U.S. participation might increase dramatically. But in fact, many people with a ballot pointed at their head simply cast a blank one or a nonsense vote for Mickey Mouse. Moreover, most countries have national elections maybe once every five years; the United States has presidential or congressional elections every two years. Frequent elections may lead to voter fatigue. New European Union elections, for instance, seem to be depressing turnout in member countries. After decades of trailing turnout in the United Kingdom, U.S. turnout in 2004 was on a par with recent British elections, in which turnout was 59.4 percent in 2001 and 61.4 percent in 2005. Americans are asked to vote more often -- in national, state, local and primary contests -- than the citizens of any other country. They can be forgiven for missing one or two elections, can't they? Even then, over the course of several elections, Americans have more chances to participate and their turnout may be higher than that in countries where people vote only once every five years. 3. Negative ads turn off voters and reduce turnout. Don't be so sure. The case on this one is still open. Negative TV advertising increased in the mid-1980s, but turnout hasn't gone down correspondingly. The negative Swift boat campaign against Sen. John F. Kerry (D-Mass.) apparently did little to depress turnout in the 2004 presidential race. Some academic studies have found that negative advertising increases turnout. And that's not so surprising: A particularly nasty ad grabs people's attention and gets them talking. People participate when they're interested. A recent GOP attack ad on Rep. Harold E. Ford Jr. (D-Tenn.), a Senate candidate, has changed the dynamic of the race, probably not because it changed minds or dissuaded Democrats, but because it energized listless Republicans. We'll have to wait to see whether the attack on Ford backfires because voters perceive it as unfair. That's the danger of going negative. So campaigns tend to stick to "contrast ads," in which candidates contrast their records with those of their opponents. When people see stark differences between candidates, they're more likely to vote. 4. The Republican "72-hour campaign" will win the election. Not necessarily. You can lead citizens to the ballot, but you can't make them vote. Republicans supposedly have a super-sophisticated last-minute get-out-the-vote effort that identifies voters who'll be pivotal in electing their candidates. Studies of a campaign's personal contact with voters through phone calls, door-to-door solicitation and the like find that it does have some positive effect on turnout. But people vote for many reasons other than meeting a campaign worker, such as the issues, the closeness of the election and the candidates' likeability. Further, these studies focus on get-out-the-vote drives in low-turnout elections, when contacts from other campaigns and outside groups are minimal. We don't know what the effects of mobilization drives are in highly competitive races in which people are bombarded by media stories, television ads and direct mail. Republican get-out-the-vote efforts could make a difference in close elections if Democrats simply sat on the sidelines. But this year Democrats have vowed to match the GOP mobilization voter for voter. So it'll take more than just knowing whether a prospective voter owns a Volvo or a BMW for Republicans to eke out victory in a competitive race. 5. Making voter registration easier would dramatically increase turnout. Well, yes and no. In 1993, the Democratic government in Washington enacted "Motor Voter," a program that allowed people to register to vote when they received their driver's license or visited a welfare office. Democrats thought that if everyone were registered, turnout rates would increase -- by as much as 7 percentage points. But while many people registered to vote, turnout didn't go up much. Subsequent studies found only small increases in turnout attributable to Motor Voter, perhaps 2 percentage points. Sizable increases in turnout can be seen in states with Election Day registration, which allows people to register when they vote. This may be related to the fact that lots of people don't make up their minds to vote until Election Day, rather than months in advance when they get a license. Authors Michael P. McDonald Publication: The Washington Post Full Article
b The Competitive Problem of Voter Turnout By webfeeds.brookings.edu Published On :: Tue, 31 Oct 2006 00:00:00 -0500 On November 7, millions of Americans will exercise their civic duty to vote. At stake will be control of the House and Senate, not to mention the success of individual candidates running for office. President Bush's "stay the course" agenda will either be enabled over the next two years by a Republican Congress or knocked off kilter by a Democratic one.With so much at stake, it is not surprising that the Pew Research Center found that 51 percent of registered voters have given a lot of thought to this November's election. This is higher than any other recent midterm election, including 44 percent in 1994, the year Republicans took control of the House. If so, turnout should better the 1994 turnout rate among eligible voters of 41 percent. There is good reason to suspect that despite the high interest, turnout will not exceed 1994. The problem is that a national poll is, well, a national poll, and does not measure attitudes of voters within states and districts. People vote when there is a reason to do so. Republican and Democratic agendas are in stark contrast on important issues, but voters also need to believe that their vote will matter in deciding who will represent them. It is here that the American electoral system is broken for many voters. Voters have little choice in most elections. In 1994, Congressional Quarterly called 98 House elections as competitive. Today, they list 51. To put it another way, we are already fairly confident of the winner in nearly 90 percent of House races. Although there is no similar tracking for state legislative offices, we know that the number of elections won by less than 60 percent of the vote has fallen since 1994. The real damage to the national turnout rate is in the large states of California and New York, which together account for 17 percent of the country's eligible voters. Neither state has a competitive Senate or Governor's election, and few competitive House or state legislative races. Compare to 1994, when Californians participated in competitive Senate and governor races the state's turnout was 5 percentage points above the national rate. The same year New York's competitive governor's race helped boost turnout a point above the national rate. Lacking stimulation from two of the largest states, turnout boosts will have to come from elsewhere. Texas has an interesting four-way governor's race that might draw from infrequent voters to the polls. Ohio's competitive Senate race and some House races might also draw voters. However, in other large states like Florida, Illinois, Michigan and Pennsylvania, turnout will suffer from largely uncompetitive statewide races. The national turnout rate will likely be less than 1994 and fall shy of 40 percent. This is not to say that turnout will be poor everywhere. Energized voters in Connecticut get to vote in an interesting Senate race and three of five Connecticut House seats are up for grabs. The problem is that turnout will be localized in these few areas of competition. The fault is not on the voters; people's lives are busy, and a rational person will abstain when their vote does not matter to the election outcome. The political parties also are sensitive to competition and focus their limited resources where elections are competitive. Television advertising and other mobilizing efforts by campaigns will only be found in competitive races. The old adage of "build it and they will come" is relevant. All but hardcore sports fans tune out a blowout. Building competitive elections -- and giving voters real choices -- will do much to increase voter turnout in American politics. There are a number of reforms on the table: redistricting to create competitive districts, campaign financing to give candidates equal resources, and even altering the electoral system to fundamentally change how a vote elects representatives. If voters want choice and a government more responsive to their needs, they should consider how these seemingly arcane election procedures have real consequences on motivating them to do the most fundamental democratic action: vote. Authors Michael P. McDonald Publication: washingtonpost.com Full Article
b Collapsible Candidates from Iowa to New Hampshire By webfeeds.brookings.edu Published On :: Wed, 09 Jan 2008 12:00:00 -0500 After his first place finish in Iowa, which was supposed to propel him to a New Hampshire victory, “change” is probably a word Barack Obama does not like as much anymore. But, his support did not really change much between these two elections. He won 38 percent of Iowa’s delegates and 36 percent of New Hampshire’s vote. It was Hillary Clinton and John McCain who were the big change candidates. What happens when a presidential candidate that does well in a primary or caucus state, does not do so well in the next? The dynamic of the presidential election can swiftly and stunningly change, as it did in New Hampshire on Tuesday. How Barack Obama wishes John Edwards showed up in New Hampshire. Edwards was awarded 30 percent of Iowa’s delegates, barely denying Clinton a second place finish. He finished a distant third in New Hampshire, receiving only 17 percent of the vote. There are strong indications that a shift among his supporters helped propel Hillary Clinton to her New Hampshire victory. According to the exit polls, Edwards did 8 percentage points worse in New Hampshire among women, while Clinton did 16 percent better. Obama’s support was virtually identical, dropping a statistically insignificant 1 percentage point. Obama’s support among young people remained strong, if slightly increasing among 18-24 and 30-39 year olds. Clinton’s support remained strong and slightly increased among those 65 and older. Edwards won Iowa’s middle-aged voters, age 40-64, but it was Clinton who decisively won this coveted age demographic in New Hampshire. And where these people were 38 percent Iowa caucus attendees, they were 54 percent of New Hampshire voters. (To understand why their turnout increased, see my analysis of Iowa’s turnout .) Moving forward, the generational war is still a strong dynamic in the Democratic race, as evident in the candidates’ speech styles following the election results. In Iowa, Clinton was flanked by the ghosts of the Clinton administration. In New Hampshire, she shared the stage with a sea of young voters. In Iowa, Obama spoke of change, a message that resonates with younger people who are not part of the establishment. In New Hampshire his slogan was a message that echoes the can-do spirit of the greatest generation, “Yes, we can!” In the days between Iowa and New Hampshire, Edwards spoke about how he wanted the election to become a two-way race. One should be careful with what one wishes for. Edwards and Clinton are vying for the same support base, that when united can defeat Obama, at least in New Hampshire. In the short-term, Obama most needs Edwards to do better so that support can continue to be divided. Among Republicans, John McCain recreated his magic of eight years ago and bounced back strong from a poor Iowa showing to win New Hampshire. The Iowa and New Hampshire electorates are so different it is difficult to compare them. In Iowa, Evangelical Christians were 60 percent of the electorate, while in New Hampshire, they were only 23 percent. Mike Huckabee’s move from first in Iowa to third in New Hampshire can be clearly attributed to the shrinking of his base. His collapse paved the way for a new winner to emerge. It is thus tempting to attribute McCain’s victory solely to the different electorates, but he still had to defeat Mitt Romney to win New Hampshire. According to the exit polls, the battle between McCain and Romney is a referendum on the Bush administration. Surprisingly, McCain, who has tried to rebuild bridges with the Bush establishment since his defeat in the 2000 presidential election, is still seen as the outsider and agent of change by voters participating in the Republican nomination process. In both Iowa and New Hampshire, McCain drew his support from those who said they are angry or dissatisfied with the Bush administration. Romney drew his support from those who said they are enthusiastic or satisfied. Not surprisingly, McCain is also drawing more support from self-described Independents and Romney from Republicans. The candidates seem to understand this dynamic, too, as they gave their speeches following the election results. In a contrived bit of acting, Romney showed up on stage without a podium and shoved a prepared speech back into his pocket (if he had needed a podium, his advance team would have provided it). He appeared relaxed, delivering his speech in a personable style reminiscent of Huckabee, who is competing with Romney for those who support Bush. But he also seemed to be reaching out to Independents with a message of change. In stark contrast, McCain delivered a carefully written, almost sedate speech designed to reassure Republicans of his conservative credentials. This three-way dynamic between Huckabee, McCain, and Romney should prove fascinating as the Republican nomination process moves forward. Where Evangelicals are strong, Huckabee should do well. Where they are not, the rules governing if Independents can or cannot participate will dictate how McCain and Romney do. And we have yet to see regional candidates like Fred Thompson have their day in the sun. And then there is Rudy Giuliani, who is lying in wait in the larger states where his name recognition should give him a significant boost over the other candidates. All of this points to an extended campaign among Republicans. Michael P. McDonald is an Associate Professor at George Mason University and a Non-Resident Senior Fellow at the Brookings Institution. He studies voter turnout and is a consultant to the national exit poll organization. Authors Michael P. McDonald Full Article
b Principles for Transparency and Public Participation in Redistricting By webfeeds.brookings.edu Published On :: Thu, 17 Jun 2010 14:21:00 -0400 Scholars from the Brookings Institution and the American Enterprise Institute are collaborating to promote transparency in redistricting. In January 2010, an advisory board of experts and representatives of good government groups was convened in order to articulate principles for transparent redistricting and to identify barriers to the public and communities who wish to create redistricting plans. This document summarizes the principles for transparency in redistricting that were identified during that meeting.Benefits of a Transparent, Participative Redistricting Process The drawing of electoral districts is among the most easily manipulated and least transparent systems in democratic governance. All too often, redistricting authorities maintain their monopoly by imposing high barriers to transparency and public participation. Increasing transparency and public participation can be a powerful counterbalance by providing the public with information similar to that which is typically only available to official decision makers, which can lead to different outcomes and better representation.Increasing transparency can empower the public to shape the representation for their communities, promote public commentary and discussion about redistricting, inform legislators and redistricting authorities which district configurations their constituents and the public support, and educate the public about the electoral process. Fostering public participation can enable the public to identify their neighborhoods and communities, promote the creation of alternative maps, and facilitate an exploration of a wide range of representational possibilities. The existence of publicly-drawn maps can provide a measuring stick against which an official plan can be compared, and promote the creation of a “market” for plans that support political fairness and community representational goals.Transparency Principles All redistricting plans should include sufficient information so the public can verify, reproduce, and evaluate a plan. Transparency thus requires that:Redistricting plans must be available in non-proprietary formats. Redistricting plans must be available in a format allowing them to be easily read and analyzed with commonly-used geographic information software. The criteria used as a basis for creating plans and individual districts must be clearly documented.Creating and evaluating redistricting plans and community boundaries requires access to demographic, geographic, community, and electoral data. Transparency thus requires that:All data necessary to create legal redistricting plans and define community boundaries must be publicly available, under a license allowing reuse of these data for non-commercial purposes. All data must be accompanied by clear documentation stating the original source, the chain of ownership (provenance), and all modifications made to it.Software systems used to generate or analyze redistricting plans can be complex, impossible to reproduce, or impossible to correctly understand without documentation. Transparency thus requires that:Software used to automatically create or improve redistricting plans must be either open-source or provide documentation sufficient for the public to replicate the results using independent software. Software used to generate reports that analyze redistricting plans must be accompanied by documentation of data, methods, and procedures sufficient for the reports to be verified by the public.Services offered to the public to create or evaluate redistricting plans and community boundaries are often opaque and subject to misinterpretation unless adequately documented. Transparency thus requires that:Software necessary to replicate the creation or analysis of redistricting plans and community boundaries produced by the service must be publicly available. The service must provide the public with the ability to make available all published redistricting plans and community boundaries in non-proprietary formats that are easily read and analyzed with commonly-used geographic information software. Services must provide documentation of any organizations providing significant contributions to their operation.Promoting Public Participation New technologies provide opportunities to broaden public participation in the redistricting process. These technologies should aim to realize the potential benefits described and be consistent with the articulated transparency principles. Redistricting is a legally and technically complex process. District creation and analysis software can encourage broad participation by: being widely accessible and easy to use; providing mapping and evaluating tools that help the public to create legal redistricting plans, as well as maps identifying local communities; be accompanied by training materials to assist the public to successfully create and evaluate legal redistricting plans and define community boundaries; have publication capabilities that allow the public to examine maps in situations where there is no access to the software; and promoting social networking and allow the public to compare, exchange and comment on both official and community-produced maps.Official Endorsement from Organizations – Americans for Redistricting Reform, Brennan Center for Justice at New York University, Campaign Legal Center, Center for Governmental Studies, Center for Voting and Democracy, Common Cause, Demos, and the League of Women Voters of the United States.Attending board members – Nancy Bekavac, Director, Scientists and Engineers for America; Derek Cressman, Western Regional Director of State Operations, Common Cause; Anthony Fairfax, President, Census Channel; Representative Mike Fortner (R), Illinois General Assembly; Karin Mac Donald, Director, Statewide Database, Berkeley Law, University of California, Berkeley; Leah Rush, Executive Director, Midwest Democracy Network; Mary Wilson, President, League of Women Voters.Editors – Micah Altman, Harvard University and the Brookings Institution; Thomas E. Mann, Brookings Institution; Michael P. McDonald, George Mason University and the Brookings Institution; Norman J. Ornstein, American Enterprise Institute.This project is funded by a grant from the Sloan Foundation to the Brookings Institution and the American Enterprise Institute. Authors Micah Altman Thomas E. MannMichael P. McDonaldNorman J. Ornstein Publication: The Brookings Institution and The American Enterprise Institute Image Source: © Lucy Nicholson / Reuters Full Article
b Pulling Back the Curtain on Redistricting By webfeeds.brookings.edu Published On :: Fri, 09 Jul 2010 11:23:00 -0400 Every 10 years — unfortunately, sometimes more frequently — legislative district lines are redrawn to balance population for demographic changes revealed by the census. What goes on is much more than a simple technical adjustment of boundaries, with ramifications that largely escape public notice.Politicians often use redistricting as an opportunity to cut unfavorable constituents and potential challengers out of their districts. Barack Obama, for example, learned the rough and tumble of redistricting politics when Rep. Bobby Rush (D-Ill.) carved Obama's Chicago home out of Rush's congressional district after losing a 2000 primary challenge to Obama, then a state senator. Incumbents can also use redistricting to move favorable constituents into their districts. Obama himself used the state legislative redistricting to extend his predominantly African American district north into a wealthy area of Illinois known as the Gold Coast. This new constituency allowed Obama to hone an effective biracial campaigning style that served him well when he ran for the U.S. Senate and the presidency. Critically, these decisions are made with little or no public input or accountability. While Arizona and California are among the few states that give the public a chance to see and participate in how the boundaries are set, by using open redistricting commissions, most states gerrymander legislative lines behind closed doors. Figures from both major parties tilt the electoral playing field so much that one party is essentially assured of winning a given district, controlling the state legislature or winning the most seats in the state's congressional delegation. In other words, the democratic process is subverted. In this system, politicians select voters rather than voters electing politicians. A 2006 Pew survey found that 70 percent of registered voters had no opinion about congressional redistricting. Among the few that expressed an opinion, some mistook the question to be about school districts rather than congressional districts. For many reasons it has been hard to fault the public. An immense amount of population data must be sifted and then assembled, much like a giant jigsaw puzzle, to ensure that districts satisfy complex federal requirements relating to equal population and the Voting Rights Act, and varying state requirements that may include compactness and respect for existing political boundaries or communities. And access to these data and the software necessary to assemble and analyze them have long been out of public reach. In the previous round of redistricting, according to a 2002 survey of authorities we conducted with our colleague Karin Mac Donald, most states did not provide any tools, facilities, dedicated assistance or software to support the public in developing redistricting plans. Many states failed to provide even minimal transparency by making data available, providing information about their plans online or accepting publicly submitted plans. Many redistricting authorities have not made firm plans to support transparency or public participation in the current round of redistricting. In the coming year, however, technological advancements will enable anyone with a Web browser and an interest in how he or she is represented to draw district maps of his or her community and state that meet the same requirements as official submissions. Under the direction of scholars at the Brookings Institution and the American Enterprise Institute, and with consultation from an array of experts in redistricting issues, we have developed a set of principles for transparency and public participation. These principles have been endorsed by an array of stakeholders, including Common Cause and the League of Women Voters of the United States. Americans will be able to participate directly in their democracy by offering plans to be compared with the politician-drawn maps. The public and even the courts will no longer have to accept that whatever is devised by politicians in the backroom. The Wizard of Oz appeared powerful because he hid behind a curtain -- until it was pulled back. The time has come to pull back the curtain on redistricting. A good place to start is by passing Rep. John Tanner's Redistricting Transparency Act, which has 38 co-sponsors from both parties. If Congress will not act, state governments can follow the lead of the few states that provide for meaningful transparency and public participation. Failure to provide for transparency and public participation should be recognized for what it is: an obviously self-serving act, placing the interests of politicians above the public interest. Authors Micah Altman Michael P. McDonald Publication: The Washington Post Image Source: © Joel Page / Reuters Full Article
b Midterm Elections 2010: Driving Forces, Likely Outcomes, Possible Consequences By webfeeds.brookings.edu Published On :: Mon, 04 Oct 2010 09:30:00 -0400 Event Information October 4, 20109:30 AM - 11:30 AM EDTFalk AuditoriumThe Brookings Institution1775 Massachusetts Ave., NWWashington, DC As the recent primary in Delaware attests, this year's midterm elections continue to offer unexpected twists and raise large questions. Will the Republicans take over the House and possibly the Senate? Or has the Republican wave ebbed? What role will President Obama play in rallying seemingly dispirited Democrats -- and what effect will reaction to the sluggish economy play in rallying Republicans? Is the Tea Party more an asset or a liability to the G.O.P.'s hopes? What effect will the inevitably narrowed partisan majorities have in the last two year's of Obama's first term? And how will contests for governorships and state legislatures around the nation affect redistricting and the shape of politics to come?On October 4, a panel of Brookings Governance Studies scholars, moderated by Senior Fellow E.J. Dionne, Jr., attempted to answer these questions. Senior Fellow Thomas Mann provided an overview. Senior Fellow Sarah Binder discussed congressional dynamics under shrunken majorities or divided government. Senior Fellow William Galston offered his views on the administration’s policy prospects during the 112th Congress. Nonresident Senior Fellow Michael McDonald addressed electoral reapportionment and redistricting around the country. Video Partisan Gridlock post-Elections?GOP Influence over Redistricting, ReapportionmentWorking Within Divided GovernmentGood Conditions for GOP in 2010 Midterms Audio Midterm Elections 2010: Driving Forces, Likely Outcomes, Possible Consequences Transcript Uncorrected Transcript (.pdf) Event Materials 20101004_midterm_elections Full Article
b Web Chat: Voter Enthusiasm, Early Voting and the Midterm Elections By webfeeds.brookings.edu Published On :: Wed, 20 Oct 2010 09:16:00 -0400 With little time remaining until the midterm elections, campaigning is intensifying and the outcome for control of Congress remains uncertain. Voter enthusiasm and turnout will be big factors in the elections, where Republicans have demonstrated a leg up in the party’s primaries. On October 20, Brookings expert Michael McDonald answered your questions about what the polls and early voting are telling us about the upcoming midterm elections, in a live web chat moderated by POLITICO Assistant Editor Seung Min Kim. McDonald, with Seth McKee, is author of "Revenge of the Moderates," in today's POLITICO.The transcript of this chat follows: 12:30 Seung Min Kim: Good afternoon, everyone! We have just under two weeks until the Nov. 2 midterm elections, and the Brookings Institution's Michael McDonald is here to answer your questions. Thanks and welcome, Michael. 12:30 [Comment From Dale Dean (Arlington): ] I was wondering from the historical record how closely early results mirror the actual results. Are there systemic distortions in early voting that are the same over many elections or do they differ with each election? 12:30 Michael McDonald: Early voting does not necessarily correspond with Election Day voting. Several data sources suggest the following: Overall, prior to 2008, more Republicans tended to vote early. In 2008, it was Democrats who voted early. We have to see 2010 will be a continuation of 2008 or a reversion to previous elections. 12:30 Michael McDonald: Another important factor is the number of early votes. For high early voting states like Oregon and Washington, essentially ALL votes will be cast early. In other states that require an excuse to vote absentee, the early voting electorate will be much smaller, and have a partisan character more similar to pre-2008. 12:31 [Comment From Katy Steinmetz: ] Are black voters going to turn out for Obama like they did in 2008? Why or why not? How big of a difference do you think this will make? 12:31 Michael McDonald: Since we started surveying, pollsters have found that midterm electorates -- compared to presidential electorates -- tend to be older, wealthier, better educated, and composed of fewer minorities. Sometimes Democrats can overcome this hurdle, as they did in 2006, of course. It would be highly unusual for African-Americans to vote at the same rate as they did in 2008. In some key races, in states with large minority populations, lowered levels of minority voting could be a critical determinant to the outcome. 12:32 [Comment From tim: ] Do the polls accurately reflect the relative turnout of Democrats, GOP and Independents?12:33 Michael McDonald: Pollsters try as best they can. They try to forecast who is likely to vote by various methods that are not consistent across polling firms. So, this is as much as art as a science. There are a number of factors that may further affect the partisan composition of polls, such as if people are interviewed by live interviewers or automatically or whether or not cell phones are interviewed. 12:34 [Comment From Katy Steinmetz: ] When Republican pundits like Karl Rove predict gains of 60 or so seats in the House, does that help or hurt them (in terms of making Republicans complacent and driving Democrats to the polls)? 12:36 Michael McDonald: One of the big questions in this election is the relative effects of enthusiasm versus voter mobilization. Republicans are hoping the enthusiasm gap will help them to victory, while Democrats are banking on their organization to GOTV. So far as I can tell, neither side has a distinct edge yet. 12:37 [Comment From Casey (DC): ] I have a question about the margin of error. Let's say candidate A has been consistently polling a point above candidate B, with a 3% margin of error. Is the fact that A has beaten B in all recent polls statistically significant, even with a margin of error? That is, wouldn't it be misleading to claim that A and B are tied (due to the margin of error) since A has been beating B consistently in the same poll, even by just a point? If they're truly tied, wouldn't we see A beating B half the time and B beating A the other half?? 12:41 Michael McDonald: To quickly review, the MoE is determined by the number of respondents to a survey, and it does not linearly decline as the number of respondents increases [it declines by a factor of 1/sqrt(# of respondents)]. Suppose you have two polls with 1,000 persons each, then. You may treat them as two polls of 2,000. So, the MoE would decline, but it may not decline as much as you might think. Further, as I describe above, different pollsters use different techniques to create likely voter screens (and many other survey issues), so the polls themselves are not entirely comparable. 12:42 Michael McDonald: As a general rule, I like averaging polls and looking at trends among the same pollster. If all the polls are moving in the same direction, I tend to believe that a trend is real and not just statistical noise. 12:43 Michael McDonald: Finally (I know a long answer!): never trust a single poll. Unfortunately, the media tend to report their poll, or a surprising poll, and disregard others. 12:43 [Comment From Jazziette Devereaux (AZ): ] Do you think that early voting can prevent voters from learning facts about candidates that are presented in the feverish last two weeks of the election? 12:44 Michael McDonald: My favorite example is a John Edwards voter who was upset in 2008 that he had cast his vote before he dropped out of the race. 12:46 Michael McDonald: Early voting has certainly changed campaign dynamics. No longer can an opponent release the October surprise the last week. Their opponent gets a chance to respond. And it makes elections more expensive since campaigns need to be active throughout the entire election period. So, there are pluses and minuses. 12:46 [Comment From Mark, Greenbelt: ] Is it your feeling that early voting favors one party over another generally, or is it all case-by-case? 12:48 Michael McDonald: Prior to 2008, more Republicans voted early. In 2010, more Democrats voted early. So, far more Democrats are voting early in 2008, so it may be that 2008 was a watershed election for early voting. Still, in a state-by-state basis, Republicans tend to do better among early voters in states that require an excuse to vote an absentee ballot (early voting rates are much lower, too!). 12:48 [Comment From Rosemarie (NH): ] How do you think negative campaigning impacts turnout? 12:50 Michael McDonald: It used to be that people thought negative campaigning decreased turnout, but since then, numerous studies have shown it increases turnout. People are apt to be interested in slowing down and watching the accident on the side of the road. The media certainly enjoy covering the most negative campaigns, too. 12:50 [Comment From Malcolm, DC: ] Do you have any stats about early voting so far, and can you draw any conclusions? 12:50 Michael McDonald: They are here. So far, over 2 million people have already voted! 12:52 [Comment From Borys Ortega: ] How do you see the Obama support base (liberals, young people, etc) in terms of enthusiasm? 12:52 Seung Min Kim: And in addition to that, it seems like the White House and Democrats are doing a lot more outreach to young voters, with the MTV/BET town halls and the large rallies at universities. Do you think that will have any effect, considering young people have a low turnout rate for midterm elections? 12:53 Michael McDonald: Since we began surveying, polls consistently show that young people, minorities, the poor and uneducated tend to vote at lower rates -- perhaps the most ironic thing about this election is that the people most affected by the economic downturn are the least likely to vote. 12:55 Michael McDonald: The Democrats need to counter the Republican enthusiasm by expanding the electorate. Their strategy is to do voter mobilization targeted at the low propensity midterm voters, like the youth. We will again have to see how effective the Democrat's mobilization will be compared to the Republican's enthusiasm. 12:55 [Comment From Rosemarie (NH): ] Has there been any correlation between the level and campaign spending (especially on advertising) and the results? 12:57 Michael McDonald: A funny statistic is that the more an incumbent spends, the worse they do. This is because they are spending to counter a threat from a viable challenger. This is why this is one of the most difficult questions to answer -- surprisingly. We do not know the marginal effect of another dollar spent because the other campaign is also spending money. 12:57 [Comment From Sally: ] There was a flap this week about Univision airing ads that seek to depress Hispanic voter turnout. How common is that practice? 12:59 Michael McDonald: Voter suppression targeted at minorities has a long and ignoble history in American politics. Generally, I think everyone should vote since democracy works best when its citizens are engaged. This particular episode may ultimately backfire since it may rile up Nevada Latinos in a campaign that has had many racial overtones. 1:00 [Comment From Drew C.: ] What's your evaluation of early vote-by-mail, vs. in-person voting? Are both being done well? 1:00 Michael McDonald: In 2008, approximately 500,000 mail ballots were rejected. These were people who thought they voted by their vote did not count. 1:02 Michael McDonald: Why does this happen? People do not follow the procedures properly -- the return the ballot in the wrong envelope, they do not sign the envelope, etc. I do like California's method of allowing voters to drop their ballots off on election day at their polling places. This allows poll workers to check that the voter followed procedures. 1:03 Michael McDonald: An advantage of in-person early voting is that these problems do not occur, and their is a chance for a voter and election administrators to fix any problems, such as a first time voter forgetting to bring mandatory ID. 1:03 [Comment From Nick, DC: ] Along the lines of what Sally was asking about, we hear a lot about voter suppression, and we also hear a lot about alleged voter fraud. Are either of them really very common? And are voting machines more subject to tampering than the old paper ballots? 1:05 Michael McDonald: Vote fraud -- someone actually intentionally casting an illegal vote -- is extremely rare. When it happens, it tend to happen among mail ballots. Although there are potentially security flaws with electronic machines, there is little evidence of tampering (of course, that may be because there is no way to check!). 1:06 [Comment From Peter G.: ] If you could make one voting reform nationwide to make the system work better, what would it be? 1:08 Michael McDonald: Universal voter registration. There is plenty of evidence that our system of requiring voters to register themselves does not work well. Just about every other advanced democracy registers their own voters. In states with Election Day registration, turnout is much higher (5 to 7 percentage points). So, not only would we increase turnout, but we would get third party organizations like the now-defunct ACORN our of the business of registering voters. 1:09 [Comment From Ben Griffiths: ] You said incumbents fare worse when they spend more. is the same true of challengers? I'm thinking this year of Sharron Angle's $14 million in Nevada. Is it even possible to spend that much in the time left? 1:10 Michael McDonald: The spending in Nevada is tremendous. Despite that likely about half the voters will have already voted by Election Day -- Nevada is a high turnout state -- I think the campaigns will continue spending to the end since the election appears to be going down to the wire. 1:11 Michael McDonald: As for your first question, there is a point where a challenger spends enough money to become viable, which triggers a response in spending from an incumbent. 1:11 [Comment From Rosemarie (NH): ] Is overall turnout higher in states that allow early voting? 1:13 Michael McDonald: I testified to the U.S. Senate that I believe the answer is yes, though the turnout effects are a modest one to two points in presidential elections. There are studies that find big turnout increases in non-presidential elections. Indeed, the very first usage of all-mail ballot elections was in local jurisdictions that needed to meet threshold turnout rates to pass local bond measures. 1:13 [Comment From Nancy: ] Which party gets the early bragging rights? 1:14 Michael McDonald: So far, Democrats have jack rabbited out of the starting line in most states where we have a clue of which party's registrants are voting early. Nevada is an interesting departure, where Democrats have a lead, but it is not as great as 2008. 1:14 [Comment From Carson P.: ] One of your Brookings colleagues - Bill Galston - has proposed the idea of mandatory voting, like they do in Australia. Could that work here? Is it a good idea? 1:15 Michael McDonald: Good luck trying to convince Americans that they will be fined if they do not vote. I do not think this is practical for the U.S., though it obviously increases turnout. 1:15 [Comment From Don: ] What are the prospects for Lisa Murkowski come election day? Do you think she has a realistic shot at beating Joe Miler? 1:16 Michael McDonald: The polls are close. I think it is anyone's game in Alaska. In fact, I wrote an op-ed with my co-author Seth McKee, which was published at Politico today.1:16 [Comment From Greg Dworkin: ] Thanks for all your hard work on this! How 'institutionalized' do you see the early vote by the parties? are they incorporating early voting as part of GOTV or are they behind in realizing so many people vote early these days? 1:19 Michael McDonald: As I document with another co-author -- Tom Schaller -- the Democrats created a strong early voting GOTV organization in 2008, and Republicans only belatedly tried to mobilize their voters to vote early. We will have to see how well Democrats will roll over this organization to 2010. Eventually, I believe the Republicans will have to build as strong as an organization. Early voting allows a party to mobilize over a longer period of time. 1:19 [Comment From Mary H. Hager, PhD: ] Please clarify polling methodology. Who is reached; who is not. The role of technology (email, telephonic, etc.) in defining the subpopulation for polling data. 1:20 Michael McDonald: That is quite a tall order for a chat :) We discuss many of these issues on Pollster -- which now has a home in the politics section of Huffington Post (I also blog at Pollster). 1:21 [Comment From Don (Ossning, NY): ] Does Christine O'Donnell have a chance in Delaware? 1:21 Michael McDonald: No. 1:21 [Comment From Geoffrey V.: ] Over the years, I've gotten the sense that campaigns are moving faster, that there are more undecided voters and that many voters don't make up their minds until the last minute. Is that supported by the data? 1:23 Michael McDonald: Well, given the tremendous increase of early voting from 20% in 2004 to 30% in 2008, it appears that many voters are making up their minds sooner, not later. Still, in a midterm election, the rule has generally been that people tend to hold their ballots longer because they do not have as much information about the candidates. It appears that this election may break that previous pattern. 1:23 [Comment From Joan: ] Do you think compromise will come back to Congress after the midterms? 1:24 Michael McDonald: No. Historically, we still have a ways to go before we reach the highest levels of polarization in our politics observed in the late 19th century. 1:24 [Comment From Al Amundson, ND: ] It seems sometimes that pollsters are "surprised" by wins. Polling is so scientific these days, and there's so much money behind it -- how often does a real surprise actually occur? 1:25 Michael McDonald: Surprises more often occur in primary elections, where the electorate is difficult to predict and information is fluid. I do not expect we will be greatly surprised by the 2010 election outcomes. 1:25 [Comment From Rosemarie (NH): ] Do you think that even with early voting, people just want to get it over with, go in to vote and make up their minds while they read the ballot? 1:27 Michael McDonald: Want the campaigns to stop bugging you? Vote early if you can. Election officials track who has a mail ballot in hand and who has voted, and they share this information with the campaigns. 1:27 [Comment From Bert C.: ] How is Sharron Angle still holding on in Nevada even after her numerous public gaffes? 1:27 Michael McDonald: The economic crisis has hit Nevada VERY hard (and I don't often write in caps!). 1:28 [Comment From Peggy: ] What role do you think the Tea Party will play in future elections? Is this a one-off movement or something more serious in American politics? 1:30 Michael McDonald: Shameless plug: see my Politico op-ed. A conservative/populist movement is nothing new to American politics. At least in the short run, I expect the tea party to continue to be influential, especially if Republicans take the House -- I do not expect they will take the Senate as of today. Victories will further embolden the activists. 1:31 Michael McDonald: Thanks to everyone for your questions. Sorry I could not answer them all! 1:31 Seung Min Kim: And that's it for today. Thanks for all the great questions as we count down the days until Election Day. And thanks to Michael for his insightful answers! Authors Michael P. McDonald Image Source: © John Gress / Reuters Full Article
b Toward Public Participation in Redistricting By webfeeds.brookings.edu Published On :: Thu, 20 Jan 2011 09:00:00 -0500 Event Information January 20, 20119:00 AM - 12:00 PM ESTFalk AuditoriumThe Brookings Institution1775 Massachusetts Ave., NWWashington, DC Register for the EventThe drawing of legislative district boundaries is among the most self-interested and least transparent systems in American democratic governance. All too often, formal redistricting authorities maintain their control by imposing high barriers to transparency and to public participation in the process. Reform advocates believe that opening that process to the public could lead to different outcomes and better representation.On January 20, Brookings hosted a briefing to review how redistricting in the 50 states will unfold in the months ahead and present a number of state-based initiatives designed to increase transparency and public participation in redistricting. Brookings Nonresident Senior Fellows Micah Altman and Michael McDonald unveiled open source mapping software which enables users to create and submit their own plans, based on current census and historical election data, to redistricting authorities and to disseminate them widely. Such alternative public maps could offer viable input to the formal redistricting process.After each presentation, participants took audience questions.Learn more about Michael McDonald's Public Mapping Project » Video Injecting Transparency into RedistrictingGOP Could Re-Draw 145 DistrictsTowards a New Political CultureCoalitions Pushing for Greater TransparencyThe Public Mapping Project Audio Toward Public Participation in RedistrictingToward Public Participation in RedistrictingToward Public Participation in Redistricting Transcript Download Micah Altman's and Michael McDonald's Presentation (.pdf)Download Tim Storey's Presentation (.pdf)Download Clare Dyer's Presentation (.pdf)Download the Redistricting Authority in the States Factsheet (.pdf)Download Micah Altman's and Michael McDonald's Report (.pdf)Download the District Builder Summary (.pdf) Event Materials 0120_redistricting_altman_mcdonald0120_redistricting_storey0120_redistricting_dyer0120_redistricting_authority0120_redistricting_report0120_redistricting_summary Full Article
b @ Brookings Podcast: The Politics and Process of Congressional Redistricting By webfeeds.brookings.edu Published On :: Fri, 28 Jan 2011 11:22:00 -0500 Now that the 2010 Census is concluded, states will begin the process of reapportionment—re-drawing voting district lines to account for population shifts. Nonresident Senior Fellow Michael McDonald says redistricting has been fraught with controversy and corruption since the nation’s early days, when the first “gerrymandered” district was drawn. Two states—Arizona and California—have instituted redistricting commissions intended to insulate the process from political shenanigans, but politicians everywhere will continue to work the system to gain electoral advantage and the best chance of re-election for themselves and their parties. Subscribe to audio and video podcasts of Brookings events and policy research » previous play pause next mute unmute @ Brookings Podcast: The Politics and Process of Congressional Redistricting 07:42 Download (Help) Get Code Brookings Right-click (ctl+click for Mac) on 'Download' and select 'save link as..' Get Code Copy and paste the embed code above to your website or blog. Video States Attempt to Reform Redistricting Audio @ Brookings Podcast: The Politics and Process of Congressional Redistricting Full Article
b Early Voting: A Live Web Chat with Michael McDonald By webfeeds.brookings.edu Published On :: Wed, 26 Sep 2012 12:30:00 -0400 Event Information September 26, 201212:30 PM - 1:00 PM EDTOnline OnlyThe Brookings Institution1775 Massachusetts Ave., NWWashington, DC Register for the EventThousands of Americans are already casting their votes in the 2012 elections through a variety of vote-by-mail and in-person balloting that allows citizens to cast their votes well in advance of November 6. From military personnel posted overseas to absentee voters, these early voting opportunities give voters the opportunity to make their voices heard even when they can’t stand in line on Election Day. However, there are pitfalls in the process. Expert Michael McDonald says that while a great deal of attention has been focused on voter fraud, the untold story is that during the last presidential election, some 400,000 absentee ballots were discarded as improperly submitted. How can early voters make sure their voices are heard? What effect will absentee and other early voting programs have in this election year? On September 26, McDonald took your questions and comments in a live web chat moderated by Vivyan Tran of POLITICO. 12:30 Vivyan Tran: Welcome everyone, let's get started. 12:30 Michael McDonald: Early voting was 30% of all votes cast in the 2008 election. My expectation is that 35% of all votes in 2012 will be cast prior to Election Day. In some states, the volume will be much higher. In the battleground state of CO, about 85% of the votes will be cast early; 70% in FL; and 45% in Ohio. What does it all mean? Hopefully I will be able to answer that question in today's chat! 12:30 Comment from JMC: At what point do you think that the in person early voters become less partisan types eager to cast their vote and more "regular folks" who would be more swayed by debate performances, TV ads, and the like? 12:30 Comment from Jason: 400,000 absentee ballots were discarded in 2008? How? 12:30 Michael McDonald: Reasons why election officials reject mail ballots: unsigned, envelope not sealed, multiple ballots in one envelope, etc. 400K rejected in 2008 does not include the higher rate of spoiled ballots that typically occur with paper mail ballots compared to electronic recording devices used in polling places. Moral: make sure you follow closely the proper procedures to cast your mail ballot! 12:31 Michael McDonald: @JMC: If they are going to vote early, most people wait until the week prior to the election. Those voting now have already made up their minds. But, the polls indicate many people have already done so, so maybe we see more early voting in 2012 as a consequence. 12:31 Comment from User: It was my understanding that absentee ballots are never counted unless the race is incredibly close in a particular state? Is that true - or do the rules for that vary by state? 12:32 Michael McDonald: No, all early votes are counted. What may not be counted, depending on state law and if the election is close enough for them to matter, are provisional ballots. 12:33 Comment from Damion: The blurb here says 400,000 early votes were discarded. Shouldn't the board of elections be reprimanded for that? Who was at fault and what consequences were there? 12:33: Michael McDonald: No, these are ballots "discarded" because people did not follow proper procedures and they must be rejected by law. 12:33 Comment from Shirley: Can you Facebook your vote in? 12:34 Michael McDonald: No. However, election officials are transmitting ballots electronically to overseas citizens and military voters. Voters must print the ballot, fill it out, sign it, scan it, and return. There are ways for these voters to verify that their ballot was received. 12:35 Comment from Karen K: What kind of impact could these discards have on the 2012 election? 12:36 Michael McDonald: Difficult to say. More Republicans vote by mail (excluding all mail ballot states). But, we don't know much about those who fail to follow the procedures. They might be less educated or elderly, and thus might counter the overall trend we see in mail balloting. Who knows? 12:37 Comment from User: This is the first I've heard of so many early votes getting discarded. Is this an issue people are addressing in a serious way? 12:38 Michael McDonald: Unfortunately, we are too focused on issues like voter fraud, which are low occurrence events, when there are many more important ways in which votes are lost in the system. Hopefully we can get the message out so fewer people disenfranchise themselves. 12:39 Comment from Anonymous: What do we know so far about absentee votes for 2012? Can we tell who they're leaning toward in specific states and how? 12:40 Michael McDonald: It's a little early :) yet. One of the major changes from 2008 is that the overseas civilian ballots -- a population that leans D -- was sent ballots much earlier this year than in 2008. We'll get a much better sense of the state of play in the two weeks prior to the election. 12:41 Michael McDonald: That said, the number of absentee ballot requests is running about the same as in 2008, if not a little higher, suggesting that the early vote will indeed be higher than in 2008, and perhaps that overall turnout will be on par with 2008, too. 12:41 Comment from Leslie: So, how can I ensure my early ballot is counted? There are so many rules and regulations, I'm never sure I've brought/filled out the paperwork. 12:42 Michael McDonald: Many states and localities allow people to check on-line the status of their ballot. Do a search for your local election official's webpage to see if that is available to you. 12:42 Comment from Daryyl: Can you define provisional ballots then? 12:44 Michael McDonald: Provisional ballots are required under federal law to allow people to vote if there is a problem with their voter registration. Election officials work after the election to resolve the situation. If you vote in-person early, then you can resolve provisional ballot situations much sooner, which is good. 12:45 Michael McDonald: Some states use provisional ballots for other purposes: e.g., for a person who does not have the required id or to manage a change in voter registration address. One of the untold stories of this cycle is that FL will manage change of reg. address through provisional ballots. OH does so, and 200K provisionals were cast in 2008. Expect 300K in FL, which may mean we will not know the outcome in FL until weeks after the election. Can you say 2000? 12:45 Comment from Mark, Greenbelt: Is early voting a new phenomenon, or is it increasing? It seems we should make it easier for people to vote when they can. 12:46 Michael McDonald: We are seeing more people vote early, particularly in states that offer the option. However, only MD changed its law from 2008 to allow in-person early voting. OH is sending absentee ballot requests to all registered voters, which is not a change in law, but a change in procedure that is expected to significantly increase early voting there. 12:47 Comment from Jennifer S. : Why do we vote on Tuesday? It seems inconvenient. Wouldn't more people vote if we did it on the weekend? Or over a period of days that offered both morning and evening hours? 12:48 Michael McDonald: We used to have early voting in the US! Back at the Founding, elections were held over several days to allow people living in remote areas to get to the courthouse (the polling place back in the day) to vote. In the mid-1840s, the federal gov't set the current single day for voting because -- what else? -- claims of vote fraud. That people could vote more than once. 12:49 Comment from Winston: What percentage of the U.S. population votes? And, if you could make one change that would increase voting in the U.S. what would be? 12:50 Michael McDonald: I also calculate turnout rates for the country for the media and academics. 62.2% of the eligible voters cast a ballot that counted in 2008. If I were to wave a magic wand, I would have election day registration. California just adopted it yesterday (but starting 2015). States with EDR have +5-7 percentage points of turnout. 12:50 Comment from Bernie S.: One of your colleagues at Brookings, Bill Galston, has suggested that we make voting mandatory, as they do in Australia. What do you think of that idea? Is it even possible here? 12:51 Michael McDonald: That will never happen in a county that values individual freedom so deeply as the US. Fun fact: a few years back, AZ voters rejected a ballot initiative to have voters entered into a lottery. 12:51 Comment from James: If early voting becomes more and more common, shouldn't candidates start campaigning earlier? 12:53 Michael McDonald: They do. In fact, you will see the presidential candidates visit battleground states that have in-person early voting at the start of the period. In 2008, you could see how early voting increased in places where Obama held rallies. 12:53 Comment from Devi P. : What are the factors that drive turnout? How do we get people to the polls? And what can you say about the "microtargeting" strategies the political parties are using to get their voters out? 12:54 Michael McDonald: One of the major ways in which elections have changed in the past decade is that campaigns now place more effort into voter contacts. Over 50% of people reported a contact in 2008. These contacts are known to increase turnout rates by upwards of 10 percentage points. Even contacts from Facebook friends seems to matter! 12:54 Comment from Wendy P, Ohio: What's your position on electronic voting? Can't every voting machine be hacked? Isn't plain old paper balloting more secure? 12:56 Michael McDonald: I went to Caltech, so I am sensitive to the potential for hacking. That said, I encourage experimentation so that we can build a better system. There are counties that do hold electronic elections! 12:56 Comment from Leslie: 400,000 seems like a lot - does this actually have impact on the electoral votes, and if so, should we be worried in this coming election that a lengthy recall may occur? 12:57 Michael McDonald: It could affect the outcome. So please spread the word through your networks. This is the #1 way in which votes are lost in the system! 12:57 Comment from JVotes: Perhaps we should microtarget with ballot issues. Many Americans seem disappointed with the two candidates we have to choose from. 12:58 Michael McDonald: Actually, ballot issues are known to increase turnout. But only a small amount in a presidential election, about 1 percentage point. People vote in the main show: the presidential election. 12:58 Michael McDonald: Interesting aside on that: early voting seems to have a small turnout effect in presidential election, but a larger effect in state and local elections. 12:58 Comment from Jaime Ravenet: Is there a reading of the new voter ID requirements (in at least the 9 most contested states) that does not constitute an "abridgment" of citizens' voting rights? 1:00 Michael McDonald: Perhaps under state constitutions. But the US Supreme Court has already ruled in favor of Indiana's id law. Still, that does not mean that lawyers will try to find some way under federal law to overturn them. TX was blocked because their law was determined to be discriminatory, per Sec. 5 of the Voting Rights Act. 1:00 Vivyan Tran: Thanks for the questions everyone, see you next week! Full Article
b The Structure of the TANF Block Grant By webfeeds.brookings.edu Published On :: The 1996 welfare reform legislation replaced the Aid to Families with Dependent Children (AFDC) program with a new Temporary Assistance for Needy Families (TANF) block grant that is very different than its predecessor. In the old AFDC program, funds were used almost entirely to provide and administer cash assistance to low-income—usually single-parent—families. The federal government… Full Article
b Welfare Reform and Beyond By webfeeds.brookings.edu Published On :: The Brookings Institution's Welfare Reform & Beyond Initiative was created to inform the critical policy debates surrounding the upcoming congressional reauthorization of the Temporary Assistance for Needy Families (TANF) program and a number of related programs that were created or dramatically altered by the 1996 landmark welfare reform legislation. The goal of the project has… Full Article
b Social Security Smörgåsbord? Lessons from Sweden’s Individual Pension Accounts By webfeeds.brookings.edu Published On :: President Bush has proposed adding optional personal accounts as one of the central elements of a major Social Security reform proposal. Although many details remain to be worked out, the proposal would allow individuals who choose to do so to divert part of the money they currently pay in Social Security taxes into individual investment… Full Article
b Bridging the Social Security Divide: Lessons From Abroad By webfeeds.brookings.edu Published On :: Executive Summary Efforts by President George W. Bush to promote major reforms in the Social Security retirement program have not led to policy change, but rather to increased polarization between the two parties. And the longer we wait to address Social Security’s long-term funding problem, the bigger and more painful the changes will need to… Full Article
b But Will It Work?: Implementation Analysis to Improve Government Performance By webfeeds.brookings.edu Published On :: Executive Summary Problems that arise in the implementation process make it less likely that policy objectives will be achieved in many government programs. Implementation problems may also damage the morale and external reputations of the agencies in charge of implementation. Although many implementation problems occur repeatedly across programs and can be predicted in advance, legislators… Full Article
b Policy Leadership and the Blame Trap: Seven Strategies for Avoiding Policy Stalemate By webfeeds.brookings.edu Published On :: Editor’s Note: This paper is part of the Governance Studies Management and Leadership Initiative. Negative messages about political opponents increasingly dominate not just election campaigns in the United States, but the policymaking process as well. And politics dominated by negative messaging (also known as blame-generating) tends to result in policy stalemate. Negative messaging is attractive… Full Article
b The President's 2015 R&D Budget: Livin' with the blues By webfeeds.brookings.edu Published On :: Fri, 28 Mar 2014 07:30:00 -0400 On March 4, President Obama submitted to Congress his 2015 budget request. Keeping with the spending cap deal agreed last December with Congress, the level of federal R&D will remain flat; and, when discounted by inflation, it is slightly lower. The requested R&D amount for 2015 is $135.4 billion, only $1.7 billion greater than 2014. If we discount from this 1.2% increase the expected inflation of 1.7% we are confronting a 0.5% decline in real terms. Reaction of the Research Community The litany of complaints has started. The President’s Science and Technology Advisor, John Holdren said to AAAS: “This budget required a lot of tough choices. All of us would have preferred more." The Association of American Universities, representing 60 top research universities, put out a statement declaring that this budget does “disappointingly little to close the nation’s innovation deficit,” so defined by the gap between the appropriate level of R&D investment and current spending. What’s more, compared to 2014, the budget request has kept funding for scientific research roughly even but it has reallocated about $250 million from basic to applied research (see Table 1). Advocates of science have voiced their discontent. Take for instance the Federation of American Societies for Experimental Biology that has called the request a “disappointment to the research community” because the President’s budget came $2.5 billion short of their recommendations. The President’s Research and Development Budget 2015 Source: OMB Budget 2015 These complaints are fully expected and even justified: each interest group must defend their share of tax-revenues. Sadly, in times of austerity, these protestations are toothless. If they were to have any traction in claiming a bigger piece of the federal discretionary pie, advocates would have to make a comparative case showing what budget lines must go down to make room for more R&D. But that line of argumentation could mean suicide for the scientific community because it would throw it into direct political contest with other interests and such contests are rarely decided by the merits of the cause but by the relative political power of interest groups. The science lobby is better off issuing innocuous hortatory pronouncements rather than picking up political fights that it cannot win. Thus, the R&D slice is to remain pegged to the size of the total budget, which is not expected to grow, in the coming years, more than bonsai. The political accident of budget constraints is bound to change the scientific enterprise from within, not only in terms of the articulation of merits—which means more precise and compelling explanations for the relative importance of disciplines and programs—but also in terms of a shrewd political contest among science factions. Authors Walter D. Valdivia Full Article
b The Study of the Distributional Outcomes of Innovation: A Book Review By webfeeds.brookings.edu Published On :: Mon, 05 Jan 2015 07:30:00 -0500 Editors Note: This post is an extended version of a previous post. Cozzens, Susan and Dhanaraj Thakur (Eds). 2014. Innovation and Inequality: Emerging technologies in an unequal world. Northampton, Massachusetts: Edward Elgar. Historically, the debate on innovation has focused on the determinants of the pace of innovation—on the premise that innovation is the driver of long-term economic growth. Analysts and policymakers have taken less interest on how innovation-based growth affects income distribution. Less attention even has received the question of how innovation affects other forms of inequality such as economic opportunity, social mobility, access to education, healthcare, and legal representation, or inequalities in exposure to insalubrious environments, be these physical (through exposure to polluted air, water, food or harmful work conditions) or social (neighborhoods ridden with violence and crime). The relation between innovation, equal political representation and the right for people to have a say in the collective decisions that affect their lives can also be added to the list of neglect. But neglect has not been universal. A small but growing group of analysts have been working for at least three decades to produce a more careful picture of the relationship between innovation and the economy. A distinguished vanguard of this group has recently published a collection of case studies that illuminates our understanding of innovation and inequality—which is the title of the book. The book is edited by Susan Cozzens and Dhanaraj Thakur. Cozzens is a professor in the School of Public Policy and Vice Provost of Academic Affairs at Georgia Tech. She has studied innovation and inequality long before inequality was a hot topic and led the group that collaborated on this book. Thakur is a faculty member of the school College of Public Service and Urban Affairs at Tennessee State University (while writing the book he taught at the University of West Indies in Jamaica). He is an original and sensible voice in the study of social dimensions of communication technologies. We’d like to highlight here three aspects of the book: the research design, the empirical focus, and the conceptual framework developed from the case studies in the book. Edited volumes are all too often a collection of disparate papers, but not in this case. This book is patently the product of a research design that probes the evolution of a set of technologies across a wide variety of national settings and, at the same time, it examines the different reactions to new technologies within specific countries. The second part of the book devotes five chapters to study five emerging technologies—recombinant insulin, genetically modified corn, mobile phones, open-source software, and tissue culture—observing the contrasts and similarities of their evolution in different national environments. In turn, part three considers the experience of eight countries, four of high income—Canada, Germany, Malta, and the U.S.—and four of medium or low income—Argentina, Costa Rica, Jamaica, and Mozambique. The stories in part three tell how these countries assimilated these diverse technologies into to their economies and policy environments. The second aspect to highlight is the deliberate choice of elements for empirical focus. First, the object of inquiry is not all of technology but a discreet set of emerging technologies gaining a specificity that would otherwise be negated if they were to handle the unwieldy concept of “technology” broadly construed. At the same time, this choice reveals the policy orientation of the book because these new entrants have just started to shape the socio-technical spaces they inhabit while the spaces of older technologies have likely ossified. Second, the study offers ample variance in terms of jurisdictions under study, i.e. countries of all income levels; a decision that makes at the same time theory construction more difficult and the test of general premises more robust.[i] We can add that the book avoids sweeping generalizations. Third, they focus on technological projects and their champions, a choice that increases the rigor of the empirical analysis. This choice, naturally, narrows the space of generality but the lessons are more precise and the conjectures are presented with according modesty. The combination of a solid design and clear empirical focus allow the reader to obtain a sense of general insight from the cases taken together that could not be derived from any individual case standing alone. Economic and technology historians have tackled the effects of technological advancement, from the steam engine to the Internet, but those lessons are not easily applicable to the present because emerging technologies intimate at a different kind of reconfiguration of economic and social structures. It is still too early to know the long-term effects of new technologies like genetically modified crops or mobile phone cash-transfers, but this book does a good job providing useful concepts that begin to form an analytical framework. In addition, the mix of country case studies subverts the disciplinary separation between the economics of innovation (devoted mostly to high-income countries) and development studies (interested in middle and low income economies). As a consequence of these selections, the reader can draw lessons that are likely to apply to technologies and countries other than the ones discussed in this book. The third aspect we would like to underscore in this review is the conceptual framework. Cozzens, Thakur and their colleagues have done a service to anyone interested in pursuing the empirical and theoretical analysis of innovation and inequality. For these authors, income distribution is only one part of the puzzle. They observe that inequalities are also part of social, ethnic, and gender cleavages in society. Frances Stewart, from Oxford University, introduced the notion of horizontal inequalities or inequalities at the social group level (for instance, across ethnic groups or genders). She developed the concept to contrast vertical inequalities or inequalities operating at the individual level (such as household income or wealth). The authors of this book borrow Stewart’s concept and pay attention to horizontal inequalities in the technologies they examine and observe that new technologies enter marketplaces that are already configured under historical forms of exclusion. A dramatic example is the lack of access to recombinant insulin in the U.S., because it is expensive and minorities are less likely to have health insurance (see Table 3.1 in p. 80).[ii] Another example is how innovation opens opportunities for entrepreneurs but closes them for women in cultures that systematically exclude women from entrepreneurial activities. Another key concept is that of complementary assets. A poignant example is the failure of recombinant insulin to reach poor patients in Mozambique who are sent home with old medicine even though insulin is subsidized by the government. The reason why doctors deny the poor the new treatment is that they don’t have the literacy and household resources (e.g. a refrigerator, a clock) necessary to preserve the shots, inject themselves periodically, and read sugar blood levels. Technologies aimed at fighting poverty require complementary assets to be already in place and in the absence of them, they fail to mitigate suffering and ultimately ameliorate inequality. Another illustration of the importance of complementary assets is given by the case of Open Source Software. This technology has a nominal price of zero; however, only individuals who have computers and the time, disposition, and resources to learn how to use open source operative systems benefit. Likewise, companies without the internal resources to adapt open software will not adopt it and remain economically tied to proprietary software. These observations lead to two critical concepts elaborated in the book: distributional boundaries and the inequalities across technological transitions. Distributional boundaries refer to the reach of the benefits of new technologies, boundaries that could be geographic (as in urban/suburban or center/periphery) or across social cleavages or incomes levels. Standard models of technological diffusion assume the entire population will gradually adopt a new technology, but in reality the authors observe several factors intervene in limiting the scope of diffusion to certain groups. The most insidious factors are monopolies that exercise sufficient control over markets to levy high prices. In these markets, the price becomes an exclusionary barrier to diffusion. This is quite evident in the case of mobile phones (see table 5.1, p. 128) where monopolies (or oligopolies) have market power to create and maintain a distributional boundary between post-pay and high-quality for middle and high income clients and pre-pay and low-quality for poor customers. This boundary renders pre-pay plans doubly regressive because the per-minute rates are higher than post-pay and phone expenses represent a far larger percentage in poor people’s income. Another example of exclusion happens in GMOs because in some countries subsistence farmers cannot afford the prices for engineering seeds; a disadvantage that compounds to their cost and health problems as they have to use more and stronger pesticides. A technological transition, as used here, is an inflection point in the adoption of a technology that re-shapes its distributional boundaries. When smart phones were introduced, a new market for second-hand or hand-down phones was created in Maputo; people who could not access the top technology get stuck with a sub-par system. By looking at tissue culture they find that “whether it provides benefits to small farmers as well as large ones depends crucially on public interventions in the lower-income countries in our study” (p. 190). In fact, farmers in Costa Rica enjoy much better protections compare to those in Jamaica and Mozambique because the governmental program created to support banana tissue culture was designed and implemented as an extension program aimed at disseminating know-how among small-farmers and not exclusively to large multinational-owned farms. When introducing the same technology, because of this different policy environment, the distributional boundaries were made much more extensive in Costa Rica. This is a book devoted to present the complexity of the innovation-inequality link. The authors are generous in their descriptions, punctilious in the analysis of their case studies, and cautious and measured in their conclusions. Readers who seek an overarching theory of inequality, a simple story, or a test of causality, are bound to be disappointed. But those readers may find the highest reward from carefully reading all the case studies presented in this book, not only because of the edifying richness of the detail herein but also because they will be invited to rethink the proper way to understand and address the problem of inequality.[iii] [i] These are clearly spelled out: “we assumed that technologies, societies, and inequalities co-evolved; that technological projects are always inherently distributional; and that the distributional aspects of individual projects and portfolios of projects are open to choice.” (p. 6) [ii] This problem has been somewhat mitigated since the Affordable Healthcare Act entered into effect. [iii] Kevin Risser contributed to this posting. Authors Walter D. Valdivia Image Source: © Akhtar Soomro / Reuters Full Article
b Innovation and manufacturing labor: a value-chain perspective By webfeeds.brookings.edu Published On :: Fri, 06 Mar 2015 00:00:00 -0500 Policies and initiatives to promote U.S. manufacturing would be well advised to take a value chain perspective of this economic sector. Currently, our economic statistics do not include pre-production services to manufacturing such as research and development or design or post-production services such as repair and maintenance or sales. Yet, manufacturing firms invest heavily in these services because they are crucial to the success of their business. In a new paper, Kate Whitefoot and Walter Valdivia offer a fresh insight into the sector’s labor composition and trends by examining employment in manufacturing from a value chain perspective. While the manufacturing sector shed millions of jobs in the 2002-2010 period—a period that included the Great Recession—employment in upstream services expanded 26 percent for market analysis, 13 percent for research and development, and 23 percent for design and technical services. Average wages for these services increased over 10 percent in that period. Going forward, this pattern is likely to be repeated. Technical occupations, particularly in upstream segments are expected to have the largest increases in employment and wages. In light of the findings, the authors offer the following recommendations: Federal manufacturing policy: Expand PCAST’s Advanced Manufacturing Partnership recommendations—specifically, for developing a national system of certifications for production skills and establishing a national apprenticeship program for skilled trades in manufacturing—to include jobs outside the factory such as those in research and development, design and technical services, and market analysis. Higher education: Institutions of higher education should consider some adjustment to their curriculum with a long view of the coming changes to high-skill occupations, particularly with respect to problem identification and the management of uncertainty in highly automated work environments. In addition, universities and colleges should disseminate information among prospect and current students about occupations where the largest gains of employment and higher wage premiums are expected. Improve national statistics: Supplement the North American Industry Classification System (NAICS) with data that permits tracking the entire value chain, including the development of a demand-based classification system. This initiative could benefit from adding survey questions to replicate the data collection of countries with a Value Added Tax—without introducing the tax, that is—allowing in this manner a more accurate estimation of the value added by each participant in a production network. Whitefoot and Valdivia stress that any collective efforts aimed at invigorating manufacturing must seize the opportunities throughout the entire value chain including upstream and downstream services to production. Downloads Download the paper Authors Kate S. WhitefootWalter D. ValdiviaGina C. Adam Image Source: © Jeff Tuttle / Reuters Full Article
b University-industry partnerships can help tackle antibiotic resistant bacteria By webfeeds.brookings.edu Published On :: Wed, 25 Mar 2015 07:30:00 -0400 An academic-industrial partnership published last January in the prestigious journal Nature the results of the development of antibiotic teixobactin. The reported work is still at an early preclinical stage but it is nevertheless good news. Over the last decades the introduction of new antibiotics has slowed down nearly to a halt and over the same period we have seen a dangerous increase in antibiotic resistant bacteria. Such is the magnitude of the problem that it has attracted the attention of the U.S. government. Accepting several recommendations presented by the President’s Council of Advisors on Science and Technology (PCAST) in their comprehensive report, the Obama Administration issued last September an Executive Order establishing an interagency Task Force for combating antibiotic resistant bacteria and directing the Secretary of Human and Health Services (HHS) to establish an Advisory Council on this matter. More recently the White House issued a strategic plan to tackle this problem. Etiology of antibiotic resistance Infectious diseases have been a major cause of morbidity and mortality from time immemorial. The early discovery of sulfa drugs in the 1930s and then antibiotics in the 1940s significantly aided the fight against these scourges. Following World War II society experienced extraordinary gains in life expectancy and overall quality of life. During that period, marked by optimism, many people presumed victory over infectious diseases. However, overuse of antibiotics and a slowdown of innovation, allowed bacteria to develop resistance at such a pace that some experts now speak of a post-antibiotic era. The problem is manifold: overuse of antibiotics, slow innovation, and bacterial evolution. The overuse of antibiotics in both humans and livestock also facilitated the emergence of antibiotic resistant bacteria. Responsibility falls to health care providers who prescribed antibiotics liberally and patients who did not complete their prescribed dosages. Acknowledging this problem, the medical community has been training physicians to avoid pressures to prescribe antibiotics for children (and their parents) with infections that are likely to be viral in origin. Educational efforts are also underway to encourage patients to complete their full course of every prescribed antibiotic and not to halt treatment when symptoms ease. The excessive use of antibiotics in food-producing animals is perhaps less manageable because it affects the bottom line of farm operations. For instance, the FDA reported that even though famers were aware of the risks, antibiotics use in feedstock increased by 16 percent from 2009 to 2012. The development of antibiotics—perhaps a more adequate term would be anti-bacterial agents—indirectly contributed to the problem by being incremental and by nearly stalling two decades ago. Many revolutionary innovations in antibiotics were introduced in a first period of development that started in the 1940s and lasted about two decades. Building upon scaffolds and mechanisms discovered theretofore, a second period of incremental development followed over three decades, through to 1990s, with roughly three new antibiotics introduced every year. High competition and little differentiations rendered antibiotics less and less profitable and over a third period covering the last 20 years pharmaceutical companies have cut development of new antibiotics down to a trickle. The misguided overuse and misuse of antibiotics together with the economics of antibiotic innovation compounded the problem taking place in nature: bacteria evolves and adapts rapidly. Current policy initiatives The PCAST report recommended federal leadership and investment to combat antibiotic-resistant bacteria in three areas: improving surveillance, increasing the longevity of current antibiotics through moderated usage, and picking up the pace of development of new antibiotics and other effective interventions. To implement this strategy PCAST suggested an oversight structure that includes a Director for National Antibiotic Resistance Policy, an interagency Task Force for Combating Antibiotic Resistance Bacteria, and an Advisory Council to be established by the HHS Secretary. PCAST also recommended increasing federal support from $450 million to $900 million for core activities such as surveillance infrastructure and development of transformative diagnostics and treatments. In addition, it proposed $800 million in funding for the Biomedical Advanced Research and Development Authority to support public-private partnerships for antibiotics development. The Obama administration took up many of these recommendations and directed their implementation with the aforementioned Executive Order. More recently, it announced a National Strategy for Combating Antibiotic Resistant Bacteria to implement the recommendations of the PCAST report. The national strategy has five pillars: First, slow the emergence and spread of resistant bacteria by decreasing the abusive usage of antibiotics in health care as well as in farm animals; second, establish national surveillance efforts that build surveillance capability across human and animal environments; third, advance development and usage of rapid and innovative diagnostics to provide more accurate care delivery and data collection; forth, seek to accelerate the invention process for new antibiotics, other therapeutics and vaccines across all stages, including basic and applied research and development; finally, emphasize the importance of international collaboration and endorse the World Health Organization Action Plan to address antimicrobial resistance. University-Industry partnerships Therefore, an important cause of our antibiotic woes seems to be driven by economic logic. On one hand, pharmaceutical companies have by and large abandoned investment in antibiotic development; competition and high substitutability have led to low prices and in their financial calculation, pharmaceutical companies cannot justify new developmental efforts. On the other hand, farmers have found the use of antibiotics highly profitable and thus have no financial incentives to halt their use. There is nevertheless a mirror explanation of a political character. The federal government allocates about $30 billion for research in medicine and health through the National Institutes of Health. The government does not seek to crowd out private research investment; rather, the goal is to fund research the private sector would not conduct because the financial return of that research is too uncertain. Economic theory prescribes government intervention to address this kind of market failure. However, it is also government policy to privatize patents to discoveries made with public monies in order to facilitate their transfer from public to private organizations. An unanticipated risk of this policy is the rebalancing of the public research portfolio to accommodate the growing demand for the kind of research that feeds into attractive market niches. The risk is that the more aligned public research and private demand become, the less research attention will be directed to medical needs without great market prospects. The development of new antibiotics seems to be just that kind of neglected medical public need. If antibiotics are unattractive to pharmaceutical companies, antibiotic development should be a research priority for the NIH. We know that it is unlikely that Congress will increase public spending for antibiotic R&D in the proportion suggested by PCAST, but the NIH could step in and rebalance its own portfolio to increase antibiotic research. Either increasing NIH funding for antibiotics or NIH rebalancing its own portfolio, are political decisions that are sure to meet organized resistance even stronger than antibiotic resistance. The second mirror explanation is that farmers have a well-organized lobby. It is no surprise that the Executive Order gingerly walks over recommendations for the farming sector and avoid any hint at an outright ban of antibiotics use, lest the administration is perceived as heavy-handed. Considering the huge magnitude of the problem, a political solution is warranted. Farmers’ cooperation in addressing this national problem will have to be traded for subsidies and other extra-market incentives that compensate for loss revenues or higher costs. The administration will do well to work out the politics with farmer associations first before they organize in strong opposition to any measure to curb antibiotic use in feedstock. Addressing this challenge adequately will thus require working out solutions to the economic and political dimensions of this problem. Public-private partnerships, including university-industry collaboration, could prove to be a useful mechanism to balance the two dimensions of the equation. The development of teixobactin mentioned above is a good example of this prescription as it resulted from collaboration between the university of Bonn Germany, Northeastern University, and Novobiotic Pharmaceutical, a start-up in Cambridge Mass. If the NIH cannot secure an increase in research funding for antibiotics development and cannot rebalance substantially its portfolio, it can at least encourage Cooperative Research and Development Agreements as well as university start-ups devoted to develop new antibiotics. In order to promote public-private and university-industry partnerships, policy coordination is advised. The nascent enterprises will be assisted greatly if the government can help them raise capital connecting them to venture funding networks or implementing a loan guarantees programs specific to antibiotics. It can also allow for an expedited FDA approval which would lessen the regulatory burden. Likewise, farmers may be convinced to discontinue the risky practice if innovation in animal husbandry can effectively replace antibiotic use. Public-private partnerships, particularly through university extension programs, could provide an adequate framework to test alternative methods, scale them up, and subsidize the transition to new sustainable practices that are not financially painful to farmers. Yikun Chi contributed to this post More TechTank content available here Authors Walter D. ValdiviaMichael S. Kinch Image Source: © Reuters Staff / Reuters Full Article
b Responsible innovation: A primer for policymakers By webfeeds.brookings.edu Published On :: Tue, 05 May 2015 00:00:00 -0400 Technical change is advancing at a breakneck speed while the institutions that govern innovative activity slog forward trying to keep pace. The lag has created a need for reform in the governance of innovation. Reformers who focus primarily on the social benefits of innovation propose to unmoor the innovative forces of the market. Conversely, those who deal mostly with innovation’s social costs wish to constrain it by introducing regulations in advance of technological developments. In this paper, Walter Valdivia and David Guston argue for a different approach to reform the governance of innovation that they call "Responsible Innovation" because it seeks to imbue in the actors of the innovation system a more robust sense of individual and collective responsibility. Responsible innovation appreciates the power of free markets in organizing innovation and realizing social expectations but is self-conscious about the social costs that markets do not internalize. At the same time, the actions it recommends do not seek to slow down innovation because they do not constrain the set of options for researchers and businesses, they expand it. Responsible innovation is not a doctrine of regulation and much less an instantiation of the precautionary principle. Innovation and society can evolve down several paths and the path forward is to some extent open to collective choice. The aim of a responsible governance of innovation is to make that choice more consonant with democratic principles. Valdivia and Guston illustrate how responsible innovation can be implemented with three practical initiatives: Industry: Incorporating values and motivations to innovation decisions that go beyond the profit motive could help industry take on a long-view of those decisions and better manage its own costs associated with liability and regulation, while reducing the social cost of negative externalities. Consequently, responsible innovation should be an integral part of corporate social responsibility, considering that the latter has already become part of the language of business, from the classroom to the board room, and that is effectively shaping, in some quarters, corporate policies and decisions. Universities and National Laboratories: Centers for Responsible Innovation, fashioned after the institutional reform of Internal Review Boards to protect human subjects in research and the Offices of Technology Transfer created to commercialize academic research, could organize existing responsible innovation efforts at university and laboratory campuses. These Centers would formalize the consideration of impacts of research proposals on legal and regulatory frameworks, economic opportunity and inequality, sustainable development and the environment, as well as ethical questions beyond the integrity of research subjects. Federal Government: Federal policy should improve its protections and support of scientific research while providing mechanisms of public accountability for research funding agencies and their contractors. Demanding a return on investment for every research grant is a misguided approach that devalues research and undermines trust between Congress and the scientific community. At the same time, scientific institutions and their advocates should improve public engagement and demonstrate their willingness and ability to be responsive to societal concerns and expectations about the public research agenda. Second, if scientific research is a public good, by definition, markets are not effective commercializing it. New mechanisms to develop practical applications from federal research with little market appeal should be introduced to counterbalance the emphasis the current technology transfer system places on research ready for the market. Third, federal innovation policy needs to be better coordinated with other federal policy, including tax, industrial, and trade policy as well as regulatory regimes. It should also improve coordination with initiatives at the local and state level to improve the outcomes of innovation for each region, state, and metro area. Downloads Download the paper Authors Walter D. ValdiviaDavid H. Guston Full Article
b NASA considers public values in its Asteroid Initiative By webfeeds.brookings.edu Published On :: Tue, 19 May 2015 07:30:00 -0400 NASA’s Asteroid Initiative encompasses efforts for the human exploration of asteroids—as well as the Asteroid Grand Challenge—to enhance asteroid detection capabilities and mitigate their threat to Earth. The human space flight portion of the initiative primarily includes the Asteroid Redirect Mission (ARM), which is a proposal to put an asteroid in orbit of the moon and send astronauts to it. The program originally contemplated two alternatives for closer study: capturing a small 10m diameter asteroid versus simply recovering a boulder from a much larger asteroid. Late in March, NASA offered an update of its plans. It has decided to retrieve a boulder from an asteroid near Earth’s orbit—candidates are the asteroids 2008 EV5, Bennu, and Itokawa—and will place the boulder on the moon’s orbit to further study it. This mission will help NASA develop a host of technical capabilities. For instance, Solar Electric Propulsion uses solar electric power to charge atoms for spacecraft propulsion—in the absence of gravity, even a modicum of force can alter the trajectory of a body in outer space. Another related capability under development is the gravity tractor, which is based on the notion that even the modest mass of a spacecraft can exert sufficient gravitational force over an asteroid to ever so slightly change its orbit. The ARM spacecraft mass could be further increased by its ability to capture a boulder from the asteroid that is steering clear of the Earth, enabling a test of how humans might prevent asteroid threats in the future. Thus, NASA will have a second test of how to deflect near-Earth objects on a hazardous trajectory. The first test, implemented as part of the Deep Impact Mission, is a kinetic impactor; that is, crashing a spacecraft on an approaching object to change its trajectory. The Asteroid Initiative is a partner of the agency’s Near Earth Object Observation (NEOO) program. The goal of this program is to discover and monitor space objects traveling on a trajectory that could pose the risk of hitting Earth with catastrophic effects. The program also seeks to develop mitigation strategies. The capabilities developed by ARM could also support other programs of NASA, such as the manned exploration of Mars. NEOO has recently enjoyed an uptick of public support. It used to be funded at about $4 million in the 1990s and in 2010 was allocated a paltry $6 million. But then, a redirection of priorities—linked to the transition from the Bush to the Obama administrations—increased funding for NEOO to about $20 million in 2012 and $40 million in 2014—and NASA is seeking $50 million for 2015. It is clear that NASA officials made a compelling case for the importance of NEOO; in fact, what they are asking seems quite a modest amount if indeed asteroids pose an existential risk to life on earth. At the same time, the instrumental importance of the program and the public funds devoted to it beg the question as to whether taxpayers should have a say in the decisions NASA is making regarding how to proceed with the program. NASA has done something remarkable to help answer this question. Last November, NASA partnered with the ECAST network (Expert and Citizen Assessment of Science and Technology) to host a citizen forum assessing the Asteroid Initiative. ECAST is a consortium of science policy and advocacy organizations which specializes in citizen deliberations on science policy. The forum consisted of a dialogue with 100 citizens in Phoenix and Boston who learned more about the asteroid initiative and then commented on various aspects of the project. The participants, who were selected to approximate the demographics of the U.S. population, were asked to assess mitigation strategies to protect against asteroids. They were introduced to four strategies: civil defense, gravity tractor, kinetic impactor, and nuclear blast deflection. As part of the deliberations, they were asked to consider the two aforementioned approaches to perform ARM. A consensus emerged about the boulder retrieval option primarily because citizens thought that option offered better prospects for developing planetary defense technologies. This preference existed despite the excitement of capturing a full asteroid, which could potentially have additional economic impacts. The participants showed interest in promoting the development of mitigation capabilities at least as much as they wanted to protect traditional NASA goals such as the advancement of science and space flight technology. This is not surprising given that concerns about doomsday should reasonably take precedence over traditional research and exploration concerns. NASA could have decided to set ARM along the path of boulder retrieval exclusively on technical merits, but having conducted a citizen forum, the agency is now able to claim that this decision is also socially robust, which is to say, is responsive to public values of consensus. In this manner, NASA has shown a promising method by which research mission federal agencies can increase their public accountability. In the same spirit of responsible research and innovation, a recent Brookings paper I authored with David Guston—who is a co-founder of ECAST—proposes a number of other innovative ways in which the innovation enterprise can be made more responsive to public values and social expectations. Kudos to NASA for being at the forefront of innovation in space exploration and public accountability. Authors Walter D. Valdivia Image Source: © Handout . / Reuters Full Article
b The politics of federal R&D: A punctuated equilibrium analysis By webfeeds.brookings.edu Published On :: Wed, 17 Jun 2015 00:00:00 -0400 The fiscal budget has become a casualty of political polarization and even functions that had enjoyed bipartisan support, like research and development (R&D), are becoming divisive issues on Capitol Hill. As a result, federal R&D is likely to grow pegged to inflation or worse, decline. With the size of the pie fixed or shrinking, requests for R&D funding increases will trigger an inter-agency zero-sum game that will play out as pointless comparisons of agencies’ merit, or worse, as a contest to attract the favor of Congress or the White House. This insidious politics will be made even more so by the growing tendency of equating public accountability with the measurement of performance. Political polarization, tight budgets, and pressure for quantifiable results threaten to undermine the sustainability of public R&D. The situation begs the question: What can federal agencies do to deal with the changing politics of federal R&D? In a new paper, Walter D. Valdivia and Benjamin Y. Clark apply punctuated equilibrium theory to examine the last four decades of federal R&D, both at the aggregate and the agency level. Valdivia and Clark observe a general upward trend driven by gradual increases. In turn, budget leaps or punctuations are few and far in between and do no appear to have lasting effects. As the politics of R&D are stirred up, federal departments and agencies are sure to find that proposing punctuations is becoming more costly and risky. Consequently, agencies will be well advised in securing stable growth in their R&D budgets in the long run rather than pushing for short term budget leaps. While appropriations history would suggest the stability of R&D spending resulted from the character of the budget politics, in the future, stability will need the stewardship of R&D champions who work to institutionalize gradualism, this time, in spite of the politics. Downloads Download the paper Authors Walter D. ValdiviaBenjamin Y. Clark Full Article
b Federal R&D: Why is defense dominant yet less talked about? By webfeeds.brookings.edu Published On :: Thu, 25 Jun 2015 07:30:00 -0400 Federal departments and agencies received just above $133 billion in R&D funds in 2013. To put that figure in perspective, World Bank data for 2013 shows that, 130 countries had a GDP below that level; U.S. R&D is larger than the entire economy of 60 percent of all countries in the world. The chart below shows how those funds are allocated among the most important federal departments and agencies in terms of R&D. Those looking at these figures for the first time may be surprised to see that the Department of Defense takes about half of the pie. It should be noted however that not all federal R&D is destined to preserve U.S. military preeminence in the world. From non-defense research, 42 percent is destined to the much-needed research conducted by the National Institutes of Health, 17 percent to the research of the Department of Energy—owner of 17 celebrated national laboratories—16 percent for space exploration, and 8 percent for understanding the natural and social worlds at a fundamental level. The balance category is only lumped together for visual display not for its importance; it includes for instance the significant work of the National Oceanic and Atmospheric Administration and the National Institute of Standards and Technology. Despite the impressive size of defense R&D, we hear little about it. While much of defense research and development is classified, in time, civilian applications find their way into mainstream commercial uses—the Internet and GPS emerged from research done at DARPA. Far more visible than defense R&D is biomedical research, clean energy research, or news about truly impressive discoveries either in distant galaxies or in the depths of our oceans. What produces this asymmetry of visibility of federal R&D work? In a recent Brookings paper, a colleague and I suggest that the answer lies in the prominence of R&D in the agencies’ accounting books. In short: How visible is R&D and how much the agency seeks to discuss it in public fora depends not on the relative importance, but on how large a portion of the agency’s budget is dedicated to R&D. From a budget perspective, we identified two types of agencies performing R&D: those agencies whose main mission is to perform research and development, and those agencies that perform many functions in addition to R&D. For the former, the share of R&D in the discretionary budget is consistently high, while for the latter group, R&D is only a small part of their total budget (see the chart below). This distinction influences how agencies will argue for their R&D money, because they will make their case on the most important uses of their budget. If agencies have a low R&D share, they will keep it mixed with other functions and programs; for instance, research efforts will be justified only as supporting the main agency mission. In turn, agencies with a high R&D share must argue for their budgets highlighting the social outcomes of their work. These include three agencies whose primary mission is research (NASA, NSF, NIH), and a fourth (DoE) where research is a significant element of its mission. There is little question that the four agencies with high R&D share produce greatly beneficial research for society. Their strategy of promoting their work publicly is not only smart budget politics but also civic and pedagogical in the sense of helping taxpayers understand that their tax dollars are well-spent. However, it is interesting to observe that other agencies may be producing research of equal social impact that flies under the public radar, mainly because those agencies prefer as a matter of good budget policy to keep a low profile for their R&D work. One interesting conclusion for institutional design from this analysis is that promoting a research agency to the level of departments of government or its director to a cabinet rank position may bring prominence to its research, not because more and better research will necessarily get done but simply because that agency will seek public recognition for their work in order to justify its budget. Likewise, placing a research agency within a larger department may help conceal and protect their R&D funding; the politics of the department will focus on its main goals and R&D would recede to a concern of secondary interest in political battles. In the Politics of Federal R&D we discuss in more detail the changing politics of budget and how R&D agencies can respond. The general strategies of concealment and self-promotion are likely to become more important for agencies to protect a steady growth of their research and development budgets. Data sources: R&D data from the American Association for the Advancement of Sciences historical trends in Federal R&D. Total non-discretionary spending by federal agency from the Office of Management and Budget. Authors Walter D. Valdivia Image Source: © Edgar Su / Reuters Full Article
b Why should I buy a new phone? Notes on the governance of innovation By webfeeds.brookings.edu Published On :: Fri, 22 Jan 2016 20:00:00 -0500 A review essay of “Governance of Socio-technical Systems: Explaining Change”, edited by Susana Borrás and Jakob Edler (Edward Elgar, 2014, 207 pages). Phasing-out a useful and profitable technology I own a Nokia 2330; it’s a small brick phone that fits comfortably in the palm of my hand. People have feelings about this: mostly, they marvel at my ability to survive without a smart-phone. Concerns go beyond my wellbeing; once a friend protested that I should be aware of the costs I impose onto my friends, for instance, by asking them for precise directions to their houses. Another suggested that I cease trying to be smarter than my phone. But my reason is simple: I don’t need a smart phone. Most of the time, I don’t even need a mobile phone. I can take and place calls from my home or my office. And who really needs a phone during their commute? Still, my device will meet an untimely end. My service provider has informed me via text message that it will phase out all 2G service and explicitly encouraged me to acquire a 3G or newer model. There is a correct if simplistic explanation for this announcement: my provider is not making enough money with my account and should I switch to a newer device, they will be able to sell me a data plan. The more accurate and more complex explanation is that my mobile device is part of a communications system that is integrated to other economic and social systems. As those other systems evolve, my device is becoming incompatible with them; my carrier has determined that I should be integrated. The system integration is easy to understand from a business perspective. My carrier may very well be able to make a profit keeping my account as is, and the accounts of the legion of elderly and low-income customers who use similar devices, and still they may not find it advantageous in the long run to allow 2G devices in their network. To understand this business strategy, we need to go back no farther than the introduction of the iPhone, which in addition to being the most marketable mobile phone set a new standard platform for mobile devices. Its introduction accelerated a trend underway in the core business of carriers: the shift from voice communication to data streaming because smart phones can support layers of overlapping services that depend on fast and reliable data transfer. These services include sophisticated log capabilities, web search, geo-location, connectivity to other devices, and more recently added bio-monitoring. All those services are part of systems of their own, so it makes perfect business sense for carriers to seamlessly integrate mobile communications with all those other systems. Still, the economic rationale explains only a fraction of the systems integration underway. The communication system of mobile telephony is also integrated with regulatory, social, and cultural systems. Consider the most mundane examples: It’s hard to imagine anyone who, having shifted from paper-and-pencil to an electronic agenda, decided to switch back afterwards. We are increasingly dependent of GPS services; while it may have once served tourists who did not wish to learn how to navigate a new city, it is now a necessity for many people who without it are lost in their home town. Not needing to remember phone numbers, the time of our next appointment, or how to go back to that restaurant we really liked, is a clear example of the integration of mobile devices into our value systems. There are coordination efforts and mutual accommodation taking place: tech designers seek to adapt to changing values and we update our values to the new conveniences of slick gadgets. Government officials are engaged in the same mutual accommodation. They are asking how many phone booths must be left in public places, how to reach more people with public service announcements, and how to provide transit information in real-time when commuters need it. At the same time, tech designers are considering all existing regulations so their devices are compliant. Communication and regulatory systems are constantly being re-integrated. The will behind systems integration The integration of technical and social systems that results from innovation demands an enormous amount of planning, effort, and conflict resolution. The people involved in this process come from all quarters of the innovation ecology, including inventors, entrepreneurs, financiers, and government officials. Each of these agents may not be able to contemplate the totality of the system integration problem but they more or less understand how their respective system must evolve so as to be compatible with interrelated systems that are themselves evolving. There is a visible willfulness in the integration task that scholars of innovation call the governance of socio-technical systems. Introducing the term governance, I should emphasize that I do not mean merely the actions of governments or the actions of entrepreneurs. Rather, I mean the effort of all agents involved in the integration and re-integration of systems triggered by innovation; I mean all the coordination and mutual accommodation of agents from interrelated systems. And there is no single vehicle to transport all the relevant information for these agents. A classic representation of markets suggests that prices carry all the relevant information agents need to make optimal decisions. But it is impossible to project this model onto innovation because, as I suggested above, it does not adhere exclusively to economic logic; cultural and political values are also at stake. The governance task is therefore fragmented into pieces and assigned to each of the participants of the socio-technical systems involved, and they cannot resolve it as a profit-maximization problem. Instead, the participants must approach governance as a problem of design where the goal could be characterized as reflexive adaptation. By adaptation I mean seeking to achieve inter-system compatibility. By reflexive I mean that each actor must realize that their actions trigger adaption measures in other systems. Thus, they cannot passively adapt but rather they must anticipate the sequence of accommodations in the interaction with other agents. This is one of the most important aspects of the governance problem, because all too often neither technical nor economic criteria will suffice; quite regularly coordination must be negotiated, which is to say, innovation entails politics. The idea of governance of socio-technical systems is daunting. How do we even begin to understand it? What kinds of modes of governance exist? What are the key dimensions to understand the integration of socio-technical systems? And perhaps more pressing, who prevails in disputes about coordination and accommodation? Fortunately, Susana Borrás, from the Copenhagen Business School, and Jakob Edler, from the University of Manchester, both distinguished professors of innovation, have collected a set of case studies that shed light on these problems in an edited volume entitled Governance of Socio-technical Change: Explaining Change. What is more, they offer a very useful conceptual framework of governance that is worth reviewing here. While this volume will be of great interest to scholars of innovation—and it is written in scholarly language—I think it has great value for policymakers, entrepreneurs, and all agents involved in a practical manner in the work of innovation. Organizing our thinking on the governance of change The first question that Borrás and Edler tackle is how to characterize the different modes of governance. They start out with a heuristic typology across the two central categories: what kinds of agents drive innovation and how the actions of these agents are coordinated. Agents can represent the state or civil society, and actions can be coordinated via dominant or non-dominant hierarchies. Change led by state actors Change led by societal actors Coordination by dominant hierarchies Traditional deference to technocratic competence: command and control. Monopolistic or oligopolistic industrial organization. Coordination by non-dominant hierarchies State agents as primus inter pares. More competitive industries with little government oversight. Source: Adapted from Borrás and Adler (2015), Table 1.2, p. 13. This typology is very useful to understand why different innovative industries have different dynamics; they are governed differently. For instance, we can readily understand why consumer software and pharmaceuticals are so at odds regarding patent law. The strict (and very necessary) regulation of drug production and commercialization coupled with the oligopolistic structure of that industry creates the need and opportunity to advocate for patent protection; which is equivalent to a government subsidy. In turn, the highly competitive environment of consumer software development and its low level of regulation foster an environment where patents hinder innovation. Government intervention is neither needed nor wanted; the industry wishes to regulate itself. This typology is also useful to understand why open source applications have gained currency much faster in the consumer segment than the contractor segment of software producers. Examples of the latter is industry specific software (e.g. to operate machinery, the stock exchange, and ATMs) or software to support national security agencies. These contractors demand proprietary software and depend on the secrecy of the source code. The software industry is not monolithic, and while highly innovative in all its segments, the innovation taking place varies greatly by its mode of governance. Furthermore, we can understand the inherent conflicts in the governance of science. In principle, scientists are led by curiosity and organize their work in a decentralized and organic fashion. In practice, most of science is driven by mission-oriented governmental agencies and is organized in a rigid hierarchical system. Consider the centrality of prestige in science and how it is awarded by peer-review; a system controlled by the top brass of each discipline. There is nearly an irreconcilable contrast between the self-image of science and its actual governance. Using the Borrás-Edler typology, we could say that scientists imagine themselves as citizens of the south-east quadrant while they really inhabit the north-west quadrant. There are practical lessons from the application of this typology to current controversies. For instance, no policy instrument such as patents can have the same effect on all innovation sectors because the effect will depend on the mode of governance of the sector. This corollary may sound intuitive, yet it really is at variance with the current terms of the debate on patent protection, where assertions of its effect on innovation, in either direction, are rarely qualified. The second question Borrás and Edler address is that of the key analytical dimensions to examine socio-technical change. To this end, they draw from an ample selection of social theories of change. First, economists and sociologists fruitfully debate the advantage of social inquiry focused on agency versus institutions. Here, the synthesis offered is reminiscent of Herbert Simon’s “bounded rationality”, where the focus turns to agent decisions constrained by institutions. Second, policy scholars as well as sociologists emphasize the engineering of change. Change can be accomplished with discreet instruments such as laws and regulations, or diffused instruments such as deliberation, political participation, and techniques of conflict resolution. Third, political scientists underscore the centrality of power in the adjudication of disputes produced by systems’ change and integration. Borrás and Edler have condensed these perspectives in an analytical framework that boils down to three clean questions: who drives change? (focus on agents bounded by institutions), how is change engineered? (focus on instrumentation), and why it is accepted by society? (focus on legitimacy). The case studies contained in this edited volume illustrate the deployment of this framework with empirical research. Standards, sustainability, incremental innovation Arthur Daemmrich (Chapter 3) tells the story of how the German chemical company BASF succeeded marketing the biodegradable polymer Ecoflex. It is worth noting the dependence of BASF on government funding to develop Ecoflex, and on the German Institute for Standardization (DIN), making a market by setting standards. With this technology, BASF capitalized on the growing demand in Germany for biodegradables, and with its intense cooperation with DIN helped establish a standard that differentiate Ecoflex from the competition. By focusing on the enterprise (the innovation agent) and its role in engineering the market for its product by setting standards that would favor them, this story reveals the process of legitimation of this new technology. In effect, the certification of DIN was accepted by agribusinesses that sought to utilize biodegradable products. If BASF is an example of innovation by standards, Allison Loconto and Marc Barbier (Chapter 4) show the strategies of governing by standards. They take the case of the International Social and Environmental Accreditation and Labelling alliance (ISEAL). ISEAL, an advocate of sustainability, positions itself as a coordinating broker among standard developing organizations by offering “credibility tools” such as codes of conduct, best practices, impact assessment methods, and assurance codes. The organization advocates what is known as the tripartite system regime (TSR) around standards. TSR is a system of checks and balances to increase the credibility of producers complying with standards. The TSR regime assigns standard-setting, certification, and accreditation of the certifiers, to separate and independent bodies. The case illustrates how producers, their associations, and broker organizations work to bestow upon standards their most valuable attribute: credibility. The authors are cautious not to conflate credibility with legitimacy, but there is no question that credibility is part of the process of legitimizing technical change. In constructing credibility, these authors focus on the third question of the framework –legitimizing innovation—and from that vantage point, they illuminate the role of actors and instruments that will guide innovations in sustainability markets. While standards are instruments of non-dominant hierarchies, the classical instrument of dominant hierarchies is regulation. David Barberá-Tomás and Jordi Molas-Gallart tell the tragic consequences of an innovation in hip-replacement prosthesis that went terribly wrong. It is estimated that about 30 thousand replaced hips failed. The FDA, under the 1976 Medical Device Act, allows incremental improvements in medical devices to go into the market after only laboratory trials, assuming that any substantive innovations have already being tested in regular clinical trials. This policy was designed as an incentive for innovation, a relief from high regulatory costs. However, the authors argue, when products have been constantly improved for a number of years after an original release, any marginal improvement comes at a higher cost or higher risk—a point they refer to as the late stage of the product life-cycle. This has tilted the balance in favor of risky improvements, as illustrated by the hip prosthesis case. The story speaks to the integration of technical and cultural systems: the policy that encourages incremental innovation may alter the way medical device companies assess the relative risk of their innovations, precisely because they focus on incremental improvements over radical ones. Returning to the analytical framework, the vantage point of regulation—instrumentation—elucidates the particular complexities and biases in agents’ decisions. Two additional case studies discuss the discontinuation of the incandescent light bulb (ILB) and the emergence of translational research, both in Western Europe. The first study, authored by Peter Stegmaier, Stefan Kuhlmann and Vincent R. Visser (Chapter 6), focuses on a relatively smooth transition. There was wide support for replacing ILBs that translated in political will and a market willing to purchase new energy efficient bulbs. In effect, the new technical system was relatively easy to re-integrate to a social system in change—public values had shifted in Europe to favor sustainable consumption—and the authors are thus able to emphasize how agents make sense of the transition. Socio-technical change does not have a unique meaning: for citizens it means living in congruence with their values; for policy makers it means accruing political capital; for entrepreneurs it means new business opportunities. The case by Etienne Vignola-Gagné, Peter Biegelbauer and Daniel Lehner (Chapter 7) offers a similar lesson about governance. My reading of their multi-site study of the implementation of translational research—a management movement that seeks to bridge laboratory and clinical work in medical research—reveals how the different agents involved make sense of this organizational innovation. Entrepreneurs see a new market niche, researchers strive for increasing the impact of their work, and public officials align their advocacy for translation with the now regular calls for rendering publicly funded research more productive. Both chapters illuminate a lesson that is as old as it is useful to remember: technological innovation is interpreted in as many ways as the number of agents that participate in it. Innovation for whom? The framework and illustrations of this book are useful for those of us interested in the governance of system integration. The typology of different modes of governance and the three vantage points from which empirical analysis can be deployed are very useful indeed. Further development of this framework should include the question of how political power is redistributed by effect of innovation and the system integration and re-integration that it triggers. The question is pressing because the outcomes of innovation vary as power structures are reinforced or debilitated by the emergence of new technologies—not to mention ongoing destabilizing forces such as social movements. Put another way, the framework should be expanded to explain in which circumstances innovation exacerbates inequality. The expanded framework should probe whether the mutual accommodation is asymmetric across socio-economic groups, which is the same as asking: are poor people asked to do more adapting to new technologies? These questions have great relevance in contemporary debates about economic and political inequality. I believe that Borrás and Edler and their colleagues have done us a great service organizing a broad but dispersed literature and offering an intuitive and comprehensive framework to study the governance of innovation. The conceptual and empirical parts of the book are instructive and I look forward to the papers that will follow testing this framework. We need to better understand the governance of socio-technical change and the dynamics of systems integration. Without a unified framework of comparison, the ongoing efforts in various disciplines will not amount to a greater understanding of the big picture. I also have a selfish reason to like this book: it helps me make sense of my carrier’s push for integrating my value system to their technical system. If I decide to adapt to a newer phone, I could readily do so because I have time and other resources. But that may not be the case for many customers of 2G devices who have neither the resources nor the inclination to learn to use more complex devices. For that reason alone, I’d argue that this sort of innovation-led systems integration could be done more democratically. Still, I could meet the decision of my carrier with indifference: when the service is disconnected, I could simply try to get by without the darn toy. Note: Thanks to Joseph Schuman for an engaging discussion of this book with me. Authors Walter D. Valdivia Image Source: © Dominic Ebenbichler / Reuters Full Article
b The fair compensation problem of geoengineering By webfeeds.brookings.edu Published On :: Tue, 23 Feb 2016 09:00:00 -0500 The promise of geoengineering is placing average global temperature under human control, and is thus considered a powerful instrument for the international community to deal with global warming. While great energy has been devoted to learning more about the natural systems that it would affect, questions of political nature have received far less consideration. Taking as a given that regional effects will be asymmetric, the nations of the world will only give their consent to deploying this technology if they can be given assurances of a fair compensation mechanism, something like an insurance policy. The question of compensation reveals that the politics of geoengineering are far more difficult than the technical aspects. What is Geoengineering? In June 1991, Mount Pinatubo exploded, throwing a massive amount of volcanic sulfate aerosols into the high skies. The resulting cloud dispersed over weeks throughout the planet and cooled its temperature on average 0.5° Celsius over the next two years. If this kind of natural phenomenon could be replicated and controlled, the possibility of engineering the Earth’s climate is then within reach. Spraying aerosols in the stratosphere is one method of solar radiation management (SRM), a class of climate engineering that focuses on increasing the albedo, i.e. reflectivity, of the planet’s atmosphere. Other SRM methods include brightening clouds by increasing their content of sea salt. A second class of geo-engineering efforts focuses on carbon removal from the atmosphere and includes carbon sequestration (burying it deep underground) and increasing land or marine vegetation. Of all these methods, SRM is appealing for its effectiveness and low costs; a recent study put the cost at about $5 to $8 billion per year.1 Not only is SRM relatively inexpensive, but we already have the technological pieces that assembled properly would inject the skies with particles that reflect sunlight back into space. For instance, a fleet of modified Boeing 747s could deliver the necessary payload. Advocates of geoengineering are not too concerned about developing the technology to effect SRM, but about its likely consequences, not only in terms of slowing global warming but the effects on regional weather. And there lies the difficult question for geoengineering: the effects of SRM are likely to be unequally distributed across nations. Here is one example of these asymmetries: Julia Pongratz and colleagues at the department of Global Ecology of the Carnegie Institution for Science estimated a net increase in yields of wheat, corn, and rice from SRM modified weather. However, the study also found a redistributive effect with equatorial countries experiencing lower yields.2 We can then expect that equatorial countries will demand fair compensation to sign on the deployment of SRM, which leads to two problems: how to calculate compensation, and how to agree on a compensation mechanism. The calculus of compensation What should be the basis for fair compensation? One view of fairness could be that, every year, all economic gains derived from SRM are pooled together and distributed evenly among the regions or countries that experience economic losses. If the system pools gains from SRM and distributes them in proportion to losses, questions about the balance will only be asked in years in which gains and losses are about the same. But if losses are far greater than the gains; then this would be a form of insurance that cannot underwrite some of the incidents it intends to cover. People will not buy such an insurance policy; which is to say, some countries will not authorize SRM deployment. In the reverse, if the pool has a large balance left after paying out compensations, then winners of SRM will demand lower compensation taxes. Further complicating the problem is the question of how to separate gains or losses that can be attributed to SRM from regional weather fluctuations. Separating the SRM effect could easily become an intractable problem because regional weather patterns are themselves affected by SRM. For instance, any year that El Niño is particularly strong, the uncertainty about the net effect of SRM will increase exponentially because it could affect the severity of the oceanic oscillation itself. Science can reduce uncertainty but only to a certain degree, because the better we understand nature, the more we understand the contingency of natural systems. We can expect better explanations of natural phenomena from science, but it would be unfair to ask science to reduce greater understanding to a hard figure that we can plug into our compensation equation. Still, greater complexity arises when separating SRM effects from policy effects at the local and regional level. Some countries will surely organize better than others to manage this change, and preparation will be a factor in determining the magnitude of gains or losses. Inherent to the problem of estimating gains and losses from SRM is the inescapable subjective element of assessing preparation. The politics of compensation Advocates of geoengineering tell us that their advocacy is not about deploying SRM; rather, it is about better understanding the scientific facts before we even consider deployment. It’s tempting to believe that the accumulating science on SRM effects would be helpful. But when we consider the factors I just described above, it is quite possible that more science will also crystalize the uncertainty about exact amounts of compensation. The calculus of gain or loss, or the difference between the reality and a counterfactual of what regions and countries will experience requires certainty, but science only yields irreducible uncertainty about nature. The epistemic problems with estimating compensation are only to be compounded by the political contestation of those numbers. Even within the scientific community, different climate models will yield different results, and since economic compensation is derived from those models’ output, we can expect a serious contestation of the objectivity of the science of SRM impact estimation. Who should formulate the equation? Who should feed the numbers into it? A sure way to alienate scientists from the peoples of the world is to ask them to assert their cognitive authority over this calculus. What’s more, other parts of the compensation equation related to regional efforts to deal with SRM effect are inherently subjective. We should not forget the politics of asserting compensation commensurate to preparation effort; countries that experience low losses may also want compensation for their efforts preparing and coping with natural disasters. Not only would a compensation equation be a sham, it would be unmanageable. Its legitimacy would always be in question. The calculus of compensation may seem a way to circumvent the impasses of politics and define fairness mathematically. Ironically, it is shot through with subjectivity; is truly a political exercise. Can we do without compensation? Technological innovations are similar to legislative acts, observed Langdon Winner.3 Technical choices of the earliest stage in technical design quickly “become strongly fixed in material equipment, economic investment, and social habit, [and] the original flexibility vanishes for all practical purposes once the initial commitments are made.” For that reason, he insisted, "the same careful attention one would give to the rules, roles, and relationships of politics must also be given to such things as the building of highways, the creation of television networks, and the tailoring of seeming insignificant features on new machines." If technological change can be thought of as legislative change, we must consider how such a momentous technology as SRM can be deployed in a manner consonant with our democratic values. Engineering the planet’s weather is nothing short of passing an amendment to Planet Earth’s Constitution. One pesky clause in that constitutional amendment is a fair compensation scheme. It seems so small a clause in comparison to the extent of the intervention, the governance of deployment and consequences, and the international commitments to be made as a condition for deployment (such as emissions mitigation and adaptation to climate change). But in the short consideration afforded here, we get a glimpse of the intractable political problem of setting up a compensation scheme. And yet, if the clause were not approved by a majority of nations, a fair compensation scheme has little hope to be consonant with democratic aspirations. 1McClellan, Justin, David W Keith, Jay Apt. 2012. Cost analysis of stratospheric albedo modification delivery systems. Environmental Research Letters 7(3): 1-8. 2Pongratz, Julia, D. B. Lobell, L. Cao, K. Caldeira. 2012. Nature Climate Change 2, 101–105. 3Winner, Langdon. 1980. Do artifacts have politics? Daedalus (109) 1: 121-136. Authors Walter D. Valdivia Image Source: © Antara Photo Agency / Reuters Full Article
b The benefits of a knives-out Democratic debate By webfeeds.brookings.edu Published On :: Thu, 20 Feb 2020 13:31:50 +0000 Stop whining about Democrats criticizing each other. The idea that Democrats attacking Democrats is a risk and an avenue that will deliver reelection to Donald Trump is nonsense. Democrats must attack each other and attack each other aggressively. Vetting presidential candidates, highlighting their weaknesses and the gaps in their record is essential to building a… Full Article
b What drove Biden’s big wins on Super Tuesday? By webfeeds.brookings.edu Published On :: Wed, 04 Mar 2020 22:59:24 +0000 Brookings Senior Fellow John Hudak looks at the results of the Super Tuesday presidential primaries and examines the factors that fueled former Vice President Joe Biden's dramatic comeback, why former Mayor Bloomberg's unlimited budget couldn't save his candidacy, and which upcoming states will be the true tests of Biden and Bernie Sanders's competing visions for… Full Article
b Bernie Sanders’s failed coalition By webfeeds.brookings.edu Published On :: Tue, 10 Mar 2020 11:00:33 +0000 Throughout Bernie Sanders’s presidential campaigns in 2016 and 2020, he promised to transform the Democratic Party and American politics. He promised a “revolution” that would resonate with a powerful group of Americans who have not normally participated in politics: young voters, liberal voters, and new voters. He believed that once his call went out and… Full Article
b Why Bernie Sanders vastly underperformed in the 2020 primary By webfeeds.brookings.edu Published On :: Fri, 20 Mar 2020 16:43:18 +0000 Senator Bernie Sanders entered the 2020 Democratic primary race with a wind at his back. With a narrow loss to Hillary Clinton in 2016 and a massive political organization, Mr. Sanders set the tone for the policy conversation in the race. Soon after announcing, the Vermont senator began raising record amounts of money, largely online… Full Article
b It is time for a Cannabis Opportunity Agenda By webfeeds.brookings.edu Published On :: Mon, 23 Mar 2020 13:49:32 +0000 The 2020 election season will be a transformative time for cannabis policy in the United States, particularly as it relates to racial and social justice. Candidates for the White House and members of Congress have put forward ideas, policy proposals, and legislation that have changed the conversation around cannabis legalization. The present-day focus on cannabis… Full Article
b In administering the COVID-19 stimulus, the president’s role model should be Joe Biden By webfeeds.brookings.edu Published On :: Tue, 07 Apr 2020 20:24:12 +0000 As America plunges into recession, Congress and President Donald Trump have approved a series of aid packages to assist businesses, the unemployed, and others impacted by COVID-19. The first three aid packages will likely be supplemented by at least a fourth package, as the nation’s leaders better understand the depth and reach of the economic… Full Article
b ‘Essential’ cannabis businesses: Strategies for regulation in a time of widespread crisis By webfeeds.brookings.edu Published On :: Sun, 19 Apr 2020 18:32:19 +0000 Most state governors and cannabis regulators were underprepared for the COVID-19 pandemic, a crisis is affecting every economic sector. But because the legal cannabis industry is relatively new in most places and still evolving everywhere, the challenges are even greater. What’s more, there is no history that could help us understand how the industry will endure the current economic situation. And so, in many… Full Article
b Five books you should read to better understand Islam By webfeeds.brookings.edu Published On :: Tue, 15 Dec 2015 11:38:00 -0500 After a recent talk about my ISIS book, one of the audience members asked, “What can I read to help me not hate Islam?” I don’t think it’s a scholar’s job to persuade others to love or hate any culture. But the question was sincere, so I suggested some books that have helped me better understand Islam. I also put the question to Twitter. Below is some of what I and others came up with. Two cautions before we dive in: First, the list is obviously not exhaustive and I’ve left out overly apologetic books—in my experience, they only increase the skeptical reader’s suspicion that she’s being suckered. Second, people on Twitter gave me great suggestions but I’ve only included those I’ve read and can vouch for: Muhammad and the Quran: Two of the best books you’ll ever read about Muhammad and the Quran are also the shortest: The Koran: A Very Short Introduction and Muhammad, both by Michael Cook. He writes with great wit and deep scholarship. Other scriptures: Most non-Muslims are unaware that Islamic scripture is more than the Quran. It includes a vast collection of words and deeds attributed to Muhammad by later authors. These scriptures are sort of like the Gospels, and Muslim scholars fight over their authenticity like Christian scholars debate about the accuracy of Matthew, Mark, Luke, and John. These extra Islamic scriptures contain most of the teachings that make modern people (Muslims included) uncomfortable about Islam. One of the world’s experts on these scriptures, Jonathan Brown, has written a terrific book about them, Misquoting Muhammad. Rumi: The medieval mystic’s poems about life and death are beautiful and moving, no matter your belief system. I loved his poems so much as an undergrad that I went on to study Middle Eastern languages just so I could read his work in the original. I’m glad I first viewed Islam through the eyes of Rumi and not a group like ISIS. Neither is solely representative of Islam but both draw heavily on its scriptures and reach such different conclusions. The Bible: Many people recommended reading the Bible to decrease hate of Islam. The nerd in me leapt to the least obvious conclusion, “Ah, good idea! Reading some of the rough stuff in the Hebrew Bible is a good way to put a kindred ancient religion like Islam in perspective.” But they meant something a little less complicated: @will_mccants @jenanmoussa Read the bible and learn to love and not to hate. :-) — Dirk Lont (@Denkkracht1) December 12, 2015 It’s a worthy perspective today no matter your faith. Authors William McCants Image Source: © David Gray / Reuters Full Article
b Amid rising fears of ISIS, Obama must reassure By webfeeds.brookings.edu Published On :: Tue, 12 Jan 2016 12:45:00 -0500 As President Obama prepares to give the final State of the Union address of his presidency tonight, he’s promised to stay away from the technocrat’s laundry list of to-do’s. Instead, he’s expected to deliver a speech that will remind his fellow citizens of their ability to “come together as one American family.” It’s going to be a tough sell, especially when the citizens are terrified of outsiders and suspicious of one another. Most of the fear and paranoia revolves around the Islamic State group. Although the group poses far less of a threat to the United States than to our allies and friends in Europe and the Middle East, it is the sum of all fears in the minds of many Americans—an immigrant, terrorist, cyber, WMD, genocidal threat rolled into one. Its name alone can be invoked to indict Obama’s national security and immigration policies—substantive criticisms are unnecessary. [T]he Islamic State group...is the sum of all fears in the minds of many Americans. Most of those fears are overblown, but the president will want to tackle them each of them in his speech if he intends to calm fears and bring people together. He’ll explain why taking in refugees is not just living up to American values but also smart counterterrorism. He’ll showcase evidence that the military campaign against the Islamic State in the Middle East is bearing fruit. He’ll reassure Americans that the Islamic State can’t plant a skilled operative into this country and remind them that the best way to stop the unskilled lone wolf shooters inspired by the Islamic State is to close gun loop holes and monitor their behavior online before they act. He’ll demonstrate his commitment to blunting Islamic State recruitment, touting changes to how the government counters the Islamic State’s appeal online and in America’s big cities. All of that is well and good, but it’s a bureaucrat’s (or think tanker’s) effort at reassuring the public. To truly succeed in mitigating America’s fears and bringing citizens together, our country’s leader has to acknowledge that their fears are real and explain what our enemies hope to gain by engendering them. While Americans’ fears may be overblown, they won’t be deflated by technocratic hot air. Authors William McCants Full Article