Paragraphs

This article argues that how the United Nations (UN) conceptualizes legitimacy is not only a matter of legalism or power politics. The UN’s conception of legitimacy also utilizes concepts, language and symbolism from the religious realm. Understanding the entanglement between political and religious concepts and the ways of their verbalization at the agential level sheds light on how legitimacy became to be acknowledged as an integral part of the UN and how it changes. At the constitutional level, the article examines phrases and ‘verbal symbols’, enshrined in the Charter of the ‘secular church’ UN. They evoke intrinsic legitimacy claims based on religious concepts and discourse such as hope and salvation. At the agential level, the article illustrates how the Secretary-General verbalizes those abstract constitutional principles of legitimacy. Religious language and symbolism in the constitutional framework and agential practice of the UN does not necessarily produce an exclusive form of legitimacy. This article shows, however, that legitimacy as nested in the UN’s constitutional setting cannot exist without religious templates because they remain a matter of a ‘cultural frame’.

All Publications button
1
Publication Type
Journal Articles
Publication Date
Journal Publisher
International Relations
Authors

 

This workshop is co-sponsored by: Comparative Literature, The Contemporary, the Department of Sociology, Taube Center for Jewish Studies, and The Europe Center

 

McClatchy Hall A

Department of Sociology

Gisèle Sapiro École des Hautes Études en Sciences Sociales; Centre National de la Recherche Scientifique
Workshops
-
Inequality has become an intractable feature of the rich industrialized democracies, despite consensus among mass publics and experts that more social and economic equality is desirable. This book examines the political dynamics underlying the “new normal” of high and rising inequality since 1980. To do so, it traces the largely unsuccessful attempts of west European governments during this period to reduce socioeconomic inequalities in health. In England, France, and Finland, three quite different countries that span the range of European political economies, governments stated their intention to reduce inequalities in health — yet in all three cases, they were largely unable or unwilling to do what it would take to achieve this goal. Lynch finds that when center-left politicians take up the issue of socioeconomic inequalities in health, they do so in response to perceived taboos against redistribution, public spending and market regulation in a neoliberal era. Reframing inequality as a matter of health, rather than of the maldistribution of political or economic resources, is at best a partial solution, however: It reshapes the policy-making environment surrounding social inequality in ways that make it more difficult to reduce either socioeconomic inequality or health inequalities. Technocratic, medicalized inequality discourses result in shifting the Overton window around inequality away from tried-and-true policy remedies for inequality, and toward complex policy levers that are far more likely to fail. In short, inequality persists despite growing awareness of the harms it creates because of the way political leaders choose to talk about it — and not only because of economic necessity or demands from the electorate.
 
 
Image
Julia Lynch

Julia Lynch is an Associate Professor of Political Science at the University of Pennsylvania.  Her research focuses on the politics of inequality and social policy in the rich democracies, particularly the countries of western Europe. She has special interests in comparative health policy and the politics of health inequalities; the politics of aging; and the relationship between party systems and political economy in western Europe. Lynch serves as an expert advisor to the World Health Organization’s European regional office on issues of health equity, and is past chair of the Health Politics and Policy section of the American Political Science Association and past treasurer of the Council for European Studies. She is editor of Socio-Economic Review, a multi-disciplinary journal focusing on analytical, political and moral questions arising at the intersection of economy and society.  At Penn, Lynch is faculty director of the Penn In Washington Program and co-director of the Penn-Temple European Studies Colloquium. Lynch holds a BA from Harvard University, a PhD from the University of California, Berkeley, and has held visiting appointments at the European University Institute, Sciences Po, and Oxford.
 
Julia Lynch Speaker University of Pennsylvania
Lectures
Paragraphs

The field of text technologies is a capacious analytical framework that focuses on all textual records throughout human history, from the earliest periods of traceable communication—perhaps as early as 60,000 BCE—to the present day. At its core, it examines the material history of communication: what constitutes a text, the purposes for which it is intended, how it functions, and the social ends that it serves.

This coursebook can be used to support any pedagogical or research activities in text technologies, the history of the book, the history of information, and textually based work in the digital humanities. Through careful explanations of the field, examinations of terminology and themes, and illustrated case studies of diverse texts—from the Cyrus cylinder to the Eagles' "Hotel California"—Elaine Treharne and Claude Willan offer a clear yet nuanced overview of how humans convey meaning. Text Technologies will enable students and teachers to generate multiple lines of inquiry into how communication—its production, form and materiality, and reception—is crucial to any interpretation of culture, history, and society.

All Publications button
1
Publication Type
Books
Publication Date
Journal Publisher
Stanford University Press
Authors
-

Seminar Recording: https://youtu.be/BGjRsO0fKds

 

About this Event: Germany plays a key role in shaping European and Western policy toward Russia.  Berlin is a leading voice within the European Union on Russian issues, and Chancellor Angela Merkel co-chairs with the French president the "Normandy" effort that seeks to broker a setttlement between Ukraine and Russia to the conflict in Donbas.  Emily Haber, the German ambassador to the United States, will join us for a conversation on how Berlin sees the Russian challenge and how the West should respond.

 

About the Speaker: Emily Margarethe Haber has been German Ambassador to the United States since June 2018. 

Immediately prior to this, Haber, a career foreign service officer, was deployed to the Federal Ministry of the Interior, serving as State Secretary overseeing security and migration at the height of the refugee crisis in Europe. In this capacity, she worked closely with the US administration on topics ranging from the fight against international terrorism to global cyberattacks and cybersecurity. In 2009, she was appointed Political Director and, in 2011, State Secretary at the Foreign Office, the first woman to hold either post. 

Emily Haber is married to Hansjörg Haber. The couple has two sons. 

Emily Margarethe Haber German Ambassador to the United States
Seminars
-

The Shorenstein Asia-Pacific Research Center cordially invites its faculty, scholars, staff, affiliates, and their families to join APARC's first International Potluck Day! Join us to celebrate the diversity of APARC through a multicultural smorgasbord of food. Bring a dish from your home country or family heritage to share with the APARC community as we take the time to mix, mingle, and celebrate the diversity that makes APARC special.

Due to current circumstances, we will be postponing this event until further notice. Thank you for your understanding.

News Type
Commentary
Date
Paragraphs

"Freedom is inseparable from human dignity," says LarryDiamond for Bertelsmann Foundation talks on "How to Fix Democracy." The crisis is “bad, deepening, accelerating,” but he suggests several steps we can take to reverse the trend, such as ranked-choice voting to tackle the two-party system, and spreading “motor voter” laws to increase the number of registered voters. Watch the video here

Hero Image
screen shot 2020 02 03 at 11 11 47 am
All News button
1
-

The research on misinformation generally and fake news specifically is vast, as is coverage in media outlets. Two questions run throughout both the academic and public discourse: what explains the spread of fake news online, and what can be done about it? While there is substantial literature on who is likely to be exposed to and share fake news, these behaviors might not signal belief or effect. Conversely, there is far less work on who is able to differentiate between true and false stories and, as a result, who is most likely to believe fake news (or, conversely, not believe true news), a question that speaks directly to Facebook’s recent “community review” approach to combating the spread of fake news on its platform.

In his talk, Professor Tucker will report on initial findings from a new collaborative project between NYU’s Center for Social Media and Politics and Stanford’s Program on Democracy and the Internet designed to fill these gaps in the scholarly literature and inform the types of policy decisions being made by Facebook. The project has enlisted both professional fact checkers and random “crowds” of close to 100 people to fact check five “fresh” articles (that have appeared in the past 24 hours) per day, four days a week, for eights week using an innovative transparent and replicable algorithm for selecting the articles for fact checking. He will report on initial observations regarding (a) individual determinants of fact checking proficiency; (b) the viability using the “wisdom of the crowds” for fact checking, including examining the tradeoffs between crafting a more accurate crowd vs. a more representative crowd and (c) results from experiments designed to assess potential policy interventions to improve crowdsourcing accuracy.

About the Speaker:

Image
Joshua Tucker
Joshua A. Tucker is Professor of Politics, affiliated Professor of Russian and Slavic Studies, and affiliated Professor of Data Science at New York University. He is the Director of NYU’s Jordan Center for Advanced Study of Russia, a co-Director of the NYU Social Media and Political Participation (SMaPP) laboratory, a co-Director of the new NYU Center for Social Media and Politics, and a co-author/editor of the award-winning politics and policy blog The Monkey Cage at The Washington Post. He serves on the advisory boards of the American National Election Study, the Comparative Study of Electoral Systems, and numerous academic journals. Originally a scholar of post-communist politics, he has more recently studied social media and politics. His research in this area has included studies on the effects of network diversity on tolerance, partisan echo chambers, online hate speech, the effects of exposure to social media on political knowledge, online networks and protest, disinformation and fake news, how authoritarian regimes respond to online opposition, and Russian bots and trolls. His research has been funded by over $8 million in grants in the past three years, including a 2019 Knight Foundation “Research on the Future of an Informed Society” grant. His most recent book is the co-authored Communism’s Shadow: Historical Legacies and Contemporary Political Attitudes (Princeton University Press, 2017), and he is the co-editor of the forthcoming edited volume Social Media and Democracy (Cambridge University Press, 2020). 

News Type
Q&As
Date
Paragraphs

A Q&A with Professor Stephen Stedman, who serves as the Secretary General of the Kofi Annan Commission on Elections and Democracy in the Digital Age.

Image
Stedman Steve
Stephen Stedman, a Senior Fellow at the Freeman Spogli Institute for International Studies (FSI) at Stanford, is the director of the Kofi Annan Commission on Elections and Democracy in the Digital Age, an initiative of the Kofi Annan Foundation. The Commission is focused on studying the effects of social media on electoral integrity and the measures needed to safeguard the democratic process.  

At the World Economic Forum in Davos, Switzerland, the Commission which includes FSI’s Nathaniel Persily, Alex Stamos, and Toomas Ilves, launched a new report, Protecting Electoral Integrity in the Digital Age. The report takes an in-depth look at the challenges faced by democracy today and makes a number of recommendations as to how best to tackle the threats posed by social media to free and fair elections. On Tuesday, February 25, professors Stedman and Persily will discuss the report’s findings and recommendations during a lunch seminar from 12-1:15 PM. To learn more and to RSVP, visit the event page.

Q: What are some of the major findings of the report? Are digital technologies a threat to democracy?

Steve Stedman: Our report suggests that social media and the Internet pose an acute threat to democracy, but probably not in the way that most people assume. Many people believe that the problem is a diffuse one based on excess disinformation and a decline in the ability of citizens to agree on facts. We too would like the quality of deliberation in our democracy to improve and we worry about how social media might degrade democratic debate, but if we are talking about existential threats to democracy the problem is that digital technologies can be weaponized to undermine the integrity of elections.

When we started our work, we were struck by how many pathologies of democracy are said to be caused by social media: political polarization; distrust in fellow citizens, government institutions and traditional media; the decline in political parties; democratic deliberation, and on and on. Social media is said to lessen the quality of democracy because it encourages echo chambers and filter bubbles where we only interact with those who share our political beliefs. Some platforms are said to encourage extremism through their algorithms.

What we found, instead, is a much more complex problem. Many of the pathologies that social media are said to create – for instance, polarization, distrust, and political sorting begin their trendlines before the invention of the Internet, let alone the smart phone. Some of the most prominent claims are unsupported by evidence, or are confounded by conflicting evidence. In fact, we say that some assertions simply cannot be judged without access to data held by the tech platforms.

Instead, we rely on the work of scholars like Yochai Benkler and Edda Humphries to argue that not all democracies are equally vulnerable to network propaganda and disinformation. It is precisely where you have high pre-existing affective polarization, low trust, and hyperpartisan media, that digital technologies can intensify and amplify polarization.

Elections and toxic polarization are a volatile mix. Weaponized disinformation and hate speech can wreak havoc on elections, even if they don’t alter the vote tallies. This is because democracies require a system of mutual security. In established democracies political candidates and followers take it for granted that if they lose an election, they will be free to organize and contest future elections. They are confident that the winners will not use their power to eliminate them or disenfranchise them. Winners have the expectation that they hold power temporarily, and accept that they cannot change the rules of competition to stay in power forever. In short, mutual security is a set of beliefs and norms that turn elections from being a one-shot game into a repeated game with a long shadow of the future.

In a situation already marred by toxic polarization, we fear that weaponized disinformation and hate speech can cause parties and followers to believe that the other side doesn’t believe in the rules of mutual security. The stakes become higher. Followers begin to believe that losing an election means losing forever. The temptation to cheat and use violence increases dramatically. 

Q: As far as political advertising, the report encourages platforms to provide more transparency about who is funding that advertising. But it also asks that platforms require candidates to make a pledge that they will avoid deceptive campaign practices when purchasing ads. It also goes as far as to recommend financial penalties for a platform if, for example, a bot spreading information is not labelled as such. Some platforms might argue that this puts an unfair onus on them. How might platforms be encouraged to participate in this effort?

SS: The platforms have a choice: they can contribute to toxic levels of political polarization and the degradation of democratic deliberation, or they can protect electoral integrity and democracy. There are a lot of employees of the platforms who are alarmed at the state of polarization in this country and don’t want their products to be conduits of weaponized disinformation and hate speech. You saw this in the letter signed by Facebook employees objecting to the decision by Mark Zuckerberg that Facebook would treat political advertising as largely exempt from their community standards. If ever there were a moment in this country that we should demand that our political parties and candidates live up to a higher ethical standard it is now. Instead Facebook decided to allow political candidates to pay to run ads even if the ads use disinformation, tell bald-faced lies, engage in hate speech, and use doctored video and audio. Their rationale is that this is all part of “the rough and tumble of politics.” In doing so, Facebook is in the contradictory position that it has hundreds of employees working to stop disinformation and hate speech in elections in Brazil and India, but is going to allow politicians and parties in the United States to buy ads that can use disinformation and hate speech.

Our recommendation gives Facebook an option that allows political advertisement in a way that need not enflame polarization and destroy mutual security among candidates and followers: 1.) Require that candidates, groups or parties who want to pay for political advertising on Facebook sign a pledge of ethical digital practices; 2.) Then use the standards to determine if an ad meets the pledge or not. If an ad uses deep fakes, if an ad grotesquely distorts the facts, if an ad out and out lies about what an opponent said or did, then Facebook would not accept the ad. Facebook can either help us raise our electoral politics out of the sewer or it can ensure that our politics drowns in it.

It’s worth pointing out that the platforms are only one actor in a many-sided problem. Weaponized disinformation is actively spread by unscrupulous politicians and parties; it is used by foreign countries to undermine electoral integrity; and it is often spread and amplified by irresponsible partisan traditional media. Fox News, for example, ran the crazy conspiracy story about Hilary Clinton running a pedophile ring out of a pizza parlor in DC. Individuals around the president, including the son of the first National Security Adviser tweeted the story. 

Q: While many of the recommendations focus on the role of platforms and governments, the report also proposes that public authorities promote digital and media literacy in schools as well as public interest programming for the general population. What might that look like? And how would that type of literacy help protect democracy? 

SS: Our report recommends digital literacy programs as a means to help build democratic resilience against weaponized disinformation. Having said that however, the details matter tremendously. Sam Wineburg at Stanford, who we cite, has extremely insightful ideas for how to teach citizens to evaluate the information they see on the Internet, but even he puts forward warnings: if done poorly digital literacy could simply increase citizen distrust of all media, good and bad; digital literacy in a highly polarized context begs the question of who will decide what is good and bad media. We say in passing that in addition to digital literacy we need to train citizens to understand biased assimilation of information. Digital literacy trains citizens to understand who is behind a piece of information and who benefits from it. But we also need to teach citizens to stand back and ask, “why am I predisposed to want to believe this piece of information?”

Q: Obviously access to data is critical for researchers and commissioners to do their work, analysis and reporting. One of the recommendations asks that public authorities compel major internet platforms to share meaningful data with academic institutions. Why is it so important for platforms and academia to share information?

SS: Some of the most important claims about the effects of social media can’t be evaluated without access to the data. One example we cite in the report is the controversy about whether YouTube’s algorithms radicalize individuals and send them down a rabbit hole of racist, nationalist content. This is a common claim and has appeared on the front pages of the New York Times. The research supporting the claim, however, is extremely thin, and other research disputes it. What we say is that we can’t adjudicate this argument unless YouTube were to share its data, so that researchers can see what the algorithm is doing. There are similar debates concerning the effects of Facebook. One of our commissioners, Nate Persily, has been at the forefront of working with Facebook to provide certified researchers with privacy protected data – Social Science One. Progress has been so slow that the researchers have lost patience. We hope that governments can step in and compel the platforms to share the data.

Q: This is one of the first reports to look at this problem in the Global South. Is the problem more or less critical there?

SS: Kofi Annan was very concerned that the debate about digital technologies and democracy was far too focused on Europe and the United States. Before Cambridge Analytica’s involvement in the United States and Brexit elections of 2016, its predecessor company had manipulated elections in Asia, Africa and the Caribbean. There is now a transnational industry in election manipulation.

What we found does not bode well for democracies in the rest of the world. The factors that make democracies vulnerable to network propaganda and weaponized disinformation are often present in the Global South: pre-existing polarization, low trust, and hyperpartisan traditional media. Many of these democracies already have a repertoire of electoral violence. 

On the other hand, we did find innovative partnerships in Indonesia and Mexico where Election Management Bodies, civil society organizations, and traditional media cooperated to fight disinformation during elections, often with success. An important recommendation of the report is that greater attention and resources are needed for such efforts to protect electoral integrity in the Global South. 

About the Commission on Elections and Democracy in the Digital Age

 As one of his last major initiatives, in 2018 Kofi Annan convened the Commission on Elections and Democracy in the Digital Age. The Commission includes members from civil society and government, the technology sector, academia and media; across the year 2019 they examined and reviewed the opportunities and challenges for electoral integrity created by technological innovations. Assisted by a small secretariat at Stanford University and the Kofi Annan Foundation, the Commission has undertaken extensive consultations and issue recommendations as to how new technologies, social media platforms and communication tools can be harnessed to engage, empower and educate voters, and to strengthen the integrity of elections. Visit  the Kofi Annan Foundation and the Commission on Elections and Democracy in the Digital Age for more on their work.

Hero Image
democracy stock image
All News button
1

Encina Hall
616 Jane Stanford Way
Stanford, CA  94305

0
Visiting Postdoctoral Scholar at The Europe Center, 2019-2020
jaqueline_bemmer_headshot_4x5.jpg

Jaqueline Bemmer is a postdoctoral researcher and lecturer at the Law Faculty of the University of Vienna and a Visiting Postdoctoral Scholar at The Europe Center. She is a Celticist, historian and legal scholar, specializing in early medieval Irish law as well as late Roman law. She earned her DPhil in History from the University of Oxford producing the first thesis on the law of pledging in early medieval Ireland, focused around the legal tract Bretha im Ḟuillemu Gell (Judgements on Pledge Interests). 

Her current monograph: ‘Poena: conceptions of pain and suffering in late Roman legal sources’ deals with a critical period of transition and multi-normativity in European legal history situated at the threshold between the fading Roman Empire in the West and the rise of Christendom and small Germanic kingdoms in early medieval Europe, examining normative approaches to punishment, criminal procedure and penal policies. She is hosted by Prof. Walter Scheidel.

Jaqueline Bemmer has taught Irish, Welsh and Roman law as well as Latin legal terminology. She is a member of The Royal Historical Society and the Irish Legal History Society, and was a Visiting Scholar at the Centre for Scots Law, University of Aberdeen. She is articles editor for the Journal of the European Society for Comparative Legal History and has most recently been selected for participation in the Wallace Johnson Program at The Medieval Institute at Western Michigan University.

Subscribe to Europe