Skip to content

Faculty Bibliography

Search & Filter

  • Type:
    Categories:

  • Type:
    Categories:
    Sub-Categories:

    This chapter considers where law and bioethics intersect as to the 100-year life. It tackles two different issues. The first is an exploration of the bioethics of life extension and whether such extension is something that should be pursued. The second considers attempts to extend reproduction into the late period of life and the ethics thereof.

  • Type:
    Categories:

    How do private law institutions of developing countries differ from those of developed countries? A common view is that the legal systems of the Global South are often outdated, failed transplants of Global North models, or plagued by enforcement challenges. This book project offers a different perspective by focusing on legal innovation and adaptation in the Global South. We examine how countries in the Global South have embraced legal doctrines and solutions that deviate from approaches that currently hold the status of orthodoxy in richer countries, and pursue distinct and potentially broader public policy objectives or reflect different values, in response to conditions that are commonplace in developing countries. Our analysis points to reasons why the legacy of colonialism, limited fiscal capacity, economic dependence on richer countries and macro-economic volatility may encourage lawmakers in poor countries to develop heterodox doctrines. We explore different manifestations of legal heterodoxy across various areas of private law in a range of countries in the Global South. Recognizing legal heterodoxies in the Global South enlarges our understanding of legal experiences and possibilities, and contributes to our understanding about the driving forces and direction of legal evolution around the world.

  • Type:
    Categories:
    Sub-Categories:

  • Type:
    Categories:
    Sub-Categories:

  • Type:
    Categories:
    Sub-Categories:

    Links:

    Whether social media platforms increase well-being can be explored from multiple angles. Three empirical studies raise corresponding puzzles, with implications for valuation, choice, and well-being in general. The first finds that people are willing to pay far less to use social media platforms than they would demand to stop using them. The second finds that people lose welfare from using Facebook, and that Facebook users become more anxious and depressed, but that even after experiencing a good month without Facebook, they would demand a significant amount of money to stop using the platform for an additional month. The third finds that while many people would demand a significant amount of money to stop using Instagram and Tik Tok, they would also be willing to pay to eliminate Instagram and Tik Tok from their community. Each of the three puzzles has a plausible solution, but we do not yet know the ground truth. A reasonable conclusion is that people would demand a lot of money to be excluded from social media networks, which suggests that inclusion confers significant benefits, contingent on their existence – but that if social media networks did not exist, many users would be better off. This conclusion has broad implications; it suggests that people often spend time or money on goods whose existence they deplore. The three puzzles offer broad lessons for choice and welfare, and for how to think about their relationship.

  • Type:
    Categories:
    Sub-Categories:

    Links:

    In longer-form writing with Larry Gostin, especially on global health, I have been particularly struck by how careful he is not to lose the narrative voice, especially of the vulnerable. He truly believes that these stories are “on loan” to us, and that there is an almost holy reverence and devotion we owe to the lives of those whose stories we tell.

  • Type:
    Categories:
    Sub-Categories:

    Links:

    When religious claimants sue to protect their ability to practice their faith, they usually invoke constitutional and statutory guarantees that specifically protect religious exercise. But both historically and today, they often also invoke secular rights-like freedom of speech, freedom of the press, or equal protection of the laws. And when they succeed on such claims, they set precedents not just for fellow religious believers, but everyone. As a result, many rights we now take for granted arose from religious minorities fighting for their right to preach, proselytize, and publish their religious views. No study of the Supreme Court's free speech jurisprudence would be complete, for example, without considering the pathbreaking decisions won by Jehovah's Witnesses. But while many scholars have noted the contribution of Witnesses and other religious actors to First Amendment law, little scholarship has examined the broader phenomenon of religious individuals asserting secular rights or traced how their religiously motivated court battles have affected secular individuals in secular contexts. To begin to fill that gap, this article presents a historical account of how religious minorities inspired much of the Bill of Rights' secular freedoms and how Jehovah's Witnesses in particular secured many of those rights in court. The article also illustrates how religious minorities can lay the groundwork for secular social movements by considering how Witness cases in the 1930s and 40s provided crucial legal protections for the Black civil rights movement of the 1950s and 60s. Better understanding the role religious minorities have played in shaping our basic secular freedoms has important implications for how religious-claimant cases-both old and new-should be thought of today. In contexts as diverse as a high school coach praying after games to companies like Facebook and YouTube challenging state regulation of their content-feed decisions, litigants and courts must decide both when to resolve religious claimants' cases on secular grounds and, conversely, whether to resolve secular claimants' cases by analogizing to protections afforded to religiously motivated actors.

  • Type:
    Categories:
    Sub-Categories:

    Links:

    Food waste is a major contributor to climate change and a barrier to environmental sustainability. As such, reducing food waste is estimated to be one of the most promising strategies to reverse global warming. One way to address food waste is to implement effective policies, which requires an understanding of the impact of current policies on food waste and other relevant outcomes. The current study examined the impact of United States (U.S.) states’ date label policies on food waste and foodborne illness. We found that states with more date label restrictions had greater waste, but this effect was not significant. In addition, date label restrictions and foodborne illness were not related. This shows that current date label policies do not help to reduce food waste or improve public health. In other words, current U.S. state food waste policies do not promote sustainability. We discuss the implications for date label policy content and offer future research directions.

  • Type:
    Categories:

    Ruth Stout didn’t plow, dig, water, or weed—and now her “no-work” method is everywhere. But behind her secret to the perfect garden lay other secrets.

  • Type:
    Categories:
    Sub-Categories:

    Links:

    This Viewpoint discusses prospects for populating the regulatory landscape for health and health care AI in the coming years.

  • Type:
    Categories:
    Sub-Categories:

    Links:

  • Type:
    Categories:
    Sub-Categories:

    Links:

  • Type:
    Categories:
    Sub-Categories:

    Links:

  • Type:
    Categories:
    Sub-Categories:

    Links:

  • Type:
    Categories:
    Sub-Categories:

    The Supreme Court decisions in Meyer v. Nebraska and Pierce v. Society of Sisters left us a mixed legacy, one part blessing and one part curse. Many would agree on this but differ on which part is blessing and which part curse. In my view the blessing is the doctrine of substantive due process protecting personal liberty. The Court did not use this language in these cases, but it did rule that the states’ attempts to interfere with parent rights in those cases violated the 14th Amendment to the U.S. Constitution, and the Court made it clear that in future cases the state would bear a heavy burden to justify any such attempts. In later cases, the Court looked back to Meyer and Pierce as the foundational law when it created important reproductive freedom and relationship rights—the right to contraception and abortion, the rights to sex and marriage for same-sex couples. In my view the curse is the doctrine of parent rights, given the Court’s failure to create any comparable rights for children. More on this later, but first a word on legacy.

  • Type:
    Categories:

    Links:

    This study explores preferences for algorithmic versus human decision-making across six countries using nationally representative samples. Participants evaluated ten decision scenarios, typically involving serious risks of one or another kind, in which they choose between algorithmic or human decision-makers under varying informational conditions: baseline (no additional information), brief information about the expertise of the human decision-maker, brief information about the algorithm's data-driven foundation, and a combination of both. Across all countries, a strong majority preferred human decision-making. A brief account of the expertise of the human decision maker increased that majority percentage only modestly (by three percentage points). A brief account of the data on which the algorithm relies significantly reduced the size of the majority preferring the human decisionmaker (by eleven percentage points). When information about both the human and the algorithm was provided, the size of the majority preferring the human decisionmaker was also significantly reduced (by eight percentage points). Other variables, above all prior experience with algorithms, were correlated with increases or decreases in the size of the majority favouring human decision-maker or the algorithm. Prior experiences were significantly correlated with preferences, with positive interactions reversing the baseline preference for human decisionmakers when algorithmic information was provided. Methodological robustness was ensured through OLS-, Logit-, and Poisson regression, as well as Random Forest analyses. The findings suggest that informational interventions alone have a relatively modest effect on algorithm acceptance.

  • Type:
    Categories:

    Links:

    At defining points in American history, there have been massive shifts in constitutional understandings, doctrines, and practices. Apparently settled principles, and widely accepted approaches, are discarded as erroneous, even illegitimate, in favor of new principles and approaches. Less momentously, views that were once considered unthinkable do not quite become the law on the ground but instead come to be seen as plausible and part of the mainstream. Relatedly, Americans transform how they talk and think about their Constitution-its core commitments and underlying narratives-and those transformations change our practices. These radical, dizzying changes often trigger what we term a sense of "constitutional vertigo." Our goal is to provide a conceptual map and to describe how and why constitutional vertigo occurs. First, we ask whether theories of interpretation trigger radical change or whether desires for radical change impel people to generate new (or modify old) theories of interpretation. Second, we explore why so many people experience a form of vertigo. Third, we investigate the drivers of radical constitutional change, both the familiar bottom-up pressures from "We the People" and the less-familiar top-down approaches, where legal elites back and impose a new constitutional regime. Given that the Roberts Court's changes have triggered widespread vertigo, it is an apt moment to theorize about the phenomenon writ large.

  • Type:
    Categories:
    Sub-Categories:

    Many policies take the form of nudges, defined as liberty-preserving approaches that steer people in particular directions, but that also allow them to go their own way Some nudges attempt to correct self-control problems. Some nudges attempt to counteract unrealistic optimism. Some nudges attempt to correct present bias. Some nudges attempt to correct market failures, as when people are nudged not to emit air pollution. For every conventional market failure, there is a potential nudge. For every behavioral bias (optimistic bias, present bias, availability bias, limited attention), there is a responsive nudge. There are many misconceptions about nudges and nudging, and they are a diversion.

  • Type:
    Categories:

    Links:

    Both free markets and government regulators tend to use willingness to pay (WTP) as the measure of value of goods that people do not own, and willingness to accept (WTA) as the measure of value of goods that people do own. The challenge is WTP and WTA are not perfect proxies for the welfare effects of buying or selling goods, especially when people do not have experience with those goods. The reason for use of WTP and WTA is not that they are perfect, but that they seem to be the best and the most administrable method for capturing the relevant welfare effects. At the same time, WTP and WTA might be infected by a lack of information, by behavioral biases, and by hedonic forecasting errors (all of these might be either cured or aggravated by market processes). Challenges also arise when WTP is low because people do not have money and when distributional weights might be necessary to align WTP or WTA with welfare effects. There are questions about how to proceed when WTA is much higher than WTP for the same goods; the WTP-WTA disparity has yet to be fully understood. These questions are especially challenging when valuing environmental amenities, animal welfare, and social media.

  • Type:
    Categories:
    Sub-Categories:

    Links:

    Picture a community torn over a proposed zoning law. Some are angry, others defensive, and misunderstandings abound. On social media, they broadcast insults at one another; every nuanced perspective is reduced to a viral soundbite. Yet, when they meet face-to-face and start speaking, something changes: residents begin listening more than speaking, and people begin testing ideas together. Misunderstandings fade, and trust begins to form. By the end of their discussion, they have not only softened their hostility, but discovered actionable plans that benefit everyone. This is the kind of meaningful discourse our society desperately needs. Yet our digital platforms -- designed primarily for maximizing engagement through provocative content -- have pulled us away from these core community endeavours. As a constructive path forward, we introduce the idea of conversation networks as a basis for civic communication infrastructure that combines interoperable digital apps with the thoughtful integration of AI guided by human agency.

  • Type:
    Categories:
    Sub-Categories:

  • Type:
    Categories:

    Links:

    People often judge how embarrassing an activity or condition is on the basis of its perceived prevalence. They infer prevalence in part by considering how often they hear other people discussing it. But how often a condition is discussed is a function not only of its prevalence but also of how embarrassing it is. If people fail to take this into account, they will tend to judge embarrassing conditions as being rarer, which will accentuate their embarrassment, and, in turn, further amplify their reluctance to disclose those conditions - a "spiral" of shame and silence. We present results from two studies that support the existence of such a feedback process. The first, a cross-sectional survey study, asked respondents a series of questions about different embarrassing and non-embarrassing conditions. Respondents (1) indicated whether they had the conditions, (2) judged how embarrassing the conditions were, (3) reported whether they had disclosed, or would disclose, having the conditions to others, and (4) estimated what fraction of survey respondents had the conditions. As predicted, reports of disclosure were negatively related to judgments of embarrassment, and when embarrassment was greater, estimates of prevalence were lower, both for conditions that respondents had and for conditions they did not have. The second, an experimental study, manipulated whether people received a high or low estimate of population prevalence for 5 different conditions, and found that receiving a high prevalence estimate reduced embarrassment and increased self-reported willingness to disclose the condition to others, and vice versa.

  • Type:
    Categories:

  • Type:
    Categories:

    Links:

    The number of public firms in the United States has halved since the beginning of the twenty-first century, causing consternation among corporate and securities law regulators. The dominant explanations, often advanced by Securities and Exchange Commissioners, come from over- or under-regulation of the stock market. Private firms are displacing public ones, with legal imperatives largely explaining the sharp decline in the public firm. We challenge the implications of this thinking. While the number of firms has halved, public firms’ economic weight has not halved. To the contrary, the public firm sector has held steady for the past quarter-century by every other measure we examine, growing as fast as, or faster than, the economy: Profits and stock market capitalization have grown faster than the economy, while revenues and investment have kept up with the economy’s growth. We emphasize that, at their peak, public firm profits doubled from 1996 and public firm net income rose to make up more than 6% of the country’s GDP, much more than in 1996. This rise in profit has not been stressed in prior work and has implications about what really is happening in the public firm sector. The overall picture is more one that could upset progressive critics of the large corporation than one that should worry policymakers that the public firm is fading. The second challenge we pose is whether the changing configuration of the public firm sector is primarily due to corporate and securities law’s burdens. To explain the disappearance of 3,500 of the 7,300 firms that were publicly traded in 1996, one must explain not just the disappearance of many small firms, but the disappearance of firms at, near, or larger than, the median-sized firms of 1996. For the disappearance of those larger firms, the legal explanations seem implausible while industrial organization explanations are likely to be primary. We explore real economy changes that could readily explain the reconfiguration of the American public firm sector to one that is more profitable, more valuable, and with bigger but fewer firms. Overall, we build a better baseline for thinking about the public firm sector: There are fewer firms, but the sector is more profitable and bigger, with investment, revenue, and employment growing in line with the economy’s growth since 1996, and with that growth often coming in more concentrated industries. It is stronger, not weaker.

  • Type:
    Categories:
    Sub-Categories:

    Links:

  • Type:
    Categories:
    Sub-Categories:

    If you're injuring someone, you should stop—and pay for the damage you've caused. Why, this book asks, does this simple proposition, generally accepted, not apply to climate change? In Climate Justice, a bracing challenge to status quo thinking on the ethics of climate change, renowned author and legal scholar Cass Sunstein clearly frames what's at stake and lays out the moral imperative: When it comes to climate change, everyone must be counted equally, regardless of when they live or where they live—which means that wealthy nations, which have disproportionately benefited from greenhouse gas emissions, are obliged to help future generations and people in poor nations that are particularly vulnerable. Invoking principles of corrective justice and distributive justice, Sunstein argues that rich countries should pay for the harms they have caused and that all of us are obliged to take steps to protect future generations from serious climate-related damage. He shows how “choice engines,” informed by artificial intelligence, can enable people to save money and to reduce the harms they produce. The book casts new light on the “social cost of carbon,”—the most important number in climate change debates—and explains how intergenerational neutrality and international neutrality can help all nations, crucially the United States and China, do what must be done.

  • Type:
    Categories:

    The magazine has three golden rules: never write about writers, editors, or the magazine. On the occasion of our hundredth anniversary, we’re breaking them all.

  • Type:
    Categories:
    Sub-Categories:

  • Type:
    Categories:
    Sub-Categories:

    Links:

  • Type:
    Categories:
    Sub-Categories:

    Links:

    Merger review should reflect basic precepts of decision analysis, best practices in industrial organization economics, and teachings from related fields. Unfortunately, the analytical methods in modern merger guidelines fall short. Protocols violate standard prescriptions for information collection and decision-making, rely on a market definition paradigm that deviates significantly from core models of competitive interaction, fail to leverage central advances in understanding the efficiency consequences of mergers, and contravene or ignore fundamental dynamics relating to entry. This article elaborates correct analysis and contrasts it with that embodied in modern merger guidelines generally employed throughout the developed world, including the 2023 Merger Guidelines revision in the United States.

  • Type:
    Categories:

    Links:

  • Type:
    Categories:
    Sub-Categories:

  • Type:
    Categories:
    Sub-Categories:

    This paper examines the impact of defense counsel at first appearance (CAFA) on criminal justice outcomes using randomized control trials in two Texas counties. The study evaluates the influence of CAFA on bond amounts, pretrial release, conditions, and post-magistration outcomes such as recidivism and failure to appear. Results show that while CAFA reduces bond amounts and influences bond types in one jurisdiction, its effects on pretrial release and recidivism are limited. These findings highlight jurisdictional differences and suggest that CAFA’s impact may be more modest than previous studies indicate, underscoring the need for further research in this area.

  • Type:
    Categories:
    Sub-Categories:

    A workaround is a maneuver that seems, on its face, consistent with the formal rules, but that employs those rules in an unanticipated way to circumvent a legal obstacle. Though some workarounds are tolerated or even celebrated, workarounds (and proposed workarounds) often provoke instinctive skepticism or hostility. When, if ever, is such skepticism justified? Do workarounds raise distinctive legal or public policy concerns? This Article seeks to provide a systematic normative assessment of workarounds in American public law. We argue, first, that from a general public interest perspective, the desirability of a workaround depends primarily on the desirability of the rule that is being worked around. Put simply, workarounds will typically advance the public interest when the legal obstacle being worked around does more harm than good, while workarounds will set back the public interest when the obstacle being worked around serves an important public purpose. Other objections to workarounds—for example, that they will erode government legitimacy, weaken norms of self-restraint, undermine the credibility of government commitments, or sap energy for more substantial reforms—are either empirically implausible or relatively insignificant when compared to the first-order question of whether the obstacle being circumvented is itself in the public interest. Questions concerning the legality of workarounds raise different issues. While adjudicators who emphasize the primacy of legal text should have no intrinsic objections to workarounds as such, adjudicators who place significant weight on fidelity to the purposes or functions of legal rules (or rule systems) should embrace an anti-workaround presumption. But this presumption can and should be overcome in certain cases. Most significantly - and perhaps most controversially—we argue that the anti-workaround presumption should give way when the obstacle that the challenged workaround would sidestep is itself inconsistent with the larger purposes of the rule system. The question should not be whether the alleged workaround, viewed in isolation, is inconsistent with the purposes of the relevant rules, but whether the combination of obstacle and workaround, considered together, is more inconsistent with the purposes of those rules than the obstacle standing alone. Therefore, even strong purposivists might embrace certain workarounds—including workarounds to the Senate filibuster, the statutory debt ceiling, and the Electoral College.

  • Type:
    Categories:

  • Type:
    Categories:
    Sub-Categories:

    Links:

  • Type:
    Categories:
    Sub-Categories:

    Links:

    A great deal of work in behavioral science emphasizes that statistical predictions often outperform clinical predictions. Formulas tend to do better than people do, and algorithms tend to outperform human beings, including experts. One reason is that algorithms do not show inconsistency or “noise”; another reason is that they are often free from cognitive biases. These points have broad implications for risk assessment in domains that include health, safety, and the environment. Still, there is evidence that many people distrust algorithms and would prefer a human decisionmaker. We offer a set of preliminary findings about how a tested population chooses between a human being and an algorithm. In a simple choice between the two across diverse settings, people are about equally divided in their preference. We also find that that a significant number of people are willing to shift in favor of algorithms when they learn something about them, but also that a significant number of people are unmoved by the relevant information. These findings have implications for current findings about “algorithm aversion” and “algorithm appreciation.”

  • Type:
    Categories:
    Sub-Categories:

    Links:

    This study reports a comprehensive environmental scan of the generative AI (GenAI) infrastructure in the national network for clinical and translational science across 36 institutions supported by the CTSA Program led by the National Center for Advancing Translational Sciences (NCATS) of the National Institutes of Health (NIH) at the United States. Key findings indicate a diverse range of institutional strategies, with most organizations in the experimental phase of GenAI deployment. The results underscore the need for a more coordinated approach to GenAI governance, emphasizing collaboration among senior leaders, clinicians, information technology staff, and researchers. Our analysis reveals that 53% of institutions identified data security as a primary concern, followed by lack of clinician trust (50%) and AI bias (44%), which must be addressed to ensure the ethical and effective implementation of GenAI technologies.

  • Type:
    Categories:
    Sub-Categories:

    Links:

  • Type:
    Categories:
    Sub-Categories:

    Links:

    The field of artificial intelligence (AI) has entered a new cycle of intense opportunity, fueled by advances in deep learning, including generative AI. Applications of recent advances affect many aspects of everyday life, yet nowhere is it more important to use this technology safely, effectively, and equitably than in health and health care. Here, as part of the National Academy of Medicine’s Vital Directions for Health and Health Care: Priorities for 2025 initiative, which is designed to provide guidance on pressing health care issues for the incoming presidential administration, we describe the steps needed to achieve these goals. We focus on four strategic areas: ensuring safe, effective, and trustworthy use of AI; promotion and development of an AI-competent health care workforce; investing in AI research to support the science, practice, and delivery of health and health care; and promotion of policies and procedures to clarify AI liability and responsibilities.

  • Type:
    Categories:
    Sub-Categories:

    Links:

    In this response to a recent symposium on Common Good Constitutionalism in the American Journal of Jurisprudence, I principally take up themes related to democracy, disagreement, pluralism, and authority. I emphasize that the classical legal tradition is teleological, oriented to performance standards rather than design standards. Thus it does not attempt to prescribe an identical set of constitutional and institutional arrangements for all polities everywhere, but asks whether constitutional arrangements are ordered to the common good and (thus) compatible with natural and divine law. Subject to those conditions, political authority is natural, inevitable, inescapable, and good. The possibility of social and political disagreement is just a precondition for all law, not an objection to the classical legal framework. None of this entails judicial supremacism in any form, which the classical legal tradition squarely rejects.

  • Type:
    Categories:
    Sub-Categories:

    Links:

  • Type:
    Categories:
    Sub-Categories:

  • Type:
    Categories:
    Sub-Categories:

    Links:

  • Type:
    Categories:

    Links:

    People are said to show “algorithm aversion” when they prefer human forecasters or decision-makers to algorithms, even though algorithms generally outperform people (in forecasting accuracy and/or optimal decision-making in furtherance of a specified goal). Algorithm aversion also has “softer” forms, as when people prefer human forecasters or decision-makers to algorithms in the abstract, without having clear evidence about comparative performance. Algorithm aversion has strong implications for policy and law; it suggests that those who seek to use algorithms, such as officials in federal agencies, might face serious public resistance. Algorithm aversion is a product of diverse mechanisms, including (1) a desire for agency; (2) a negative moral or emotional reaction to judgment by algorithms; (3) a belief that certain human experts have unique knowledge, unlikely to be held or used by algorithms; (4) ignorance about why algorithms perform well; and (5) asymmetrical forgiveness, or a larger negative reaction to algorithmic error than to human error. An understanding of the various mechanisms provides some clues about how to overcome algorithm aversion, and also of its boundary conditions. These clues bear on the numerous decisions in law and policy, including those of federal agencies (such as the Department of Homeland Security and the Internal Revenue Service) and those involved in the criminal justice system (such as those thinking about using algorithms for bail decisions).

  • Type:
    Categories:

    Chapter 11 was widely viewed as a failure in the first decade of the Bankruptcy Code’s operation, the 1980s. Large firms were mired in bankruptcy for years; the process was seen as expensive, inaccurate, and subject to abuse. While basic bankruptcy still has its critics and few would say it works perfectly, the contrast with bankruptcy today is stark: bankruptcies that took years in the 1980s take months in the 2020s. Multiple changes explain bankruptcy’s success—creditor learning, statutory reform, better judging and lawyering, new techniques, fuller integration of the improved mechanisms that the 1978 Code added—and we do not challenge their relevance. But in our analysis, one major change is missing from the current understanding of bankruptcy’s success: bankruptcy courts and practice in the 1980s rejected market value; today bankruptcy courts and practice accept and use market value. This shift is a major explanation for bankruptcy’s success. It reduces opportunities for conflict in bankruptcy. It speeds up proceedings. It allows firms to be repositioned in market transactions. Deals among claimants and interests are more readily reached and the firm can ride through bankruptcy without the bankruptcy process materially scarring the enterprise. This switch to market values has multiple channels: more whole-firm sales, wider and deeper access to financing for bidders and bankrupts, growing judicial deference to market valuations, and a bigger and more sophisticated private equity and distressed debt industry that buys bankrupt companies and their securities. We argue that valuation improvements explain much of the increased speed and efficiency of Chapter 11 practice over the decades. We provide evidence that valuation conflicts narrowed and that the corporate reorganization process accelerated. This market-based-valuation result has implications for bankruptcy law reform around the world. Several European and Asian nations have looked to Chapter 11 to model their own restructuring laws. We urge caution. Chapter 11 works best in conjunction with institutions that facilitate market valuation and market transactions. The United States developed such institutions only in recent decades; many nations have not developed them yet. Chapter 11 went from being viewed by many as a deficient legal structure in the 1980s to a substantial success story by the twenty-first century. The switch to market thinking across the bankruptcy spectrum—in bankruptcy transactions, in judging, and in lawyering—goes far in explaining why.

  • Type:
    Categories:
    Sub-Categories:

  • Type:
    Categories:
    Sub-Categories:

    Does Congress have authority to pass legislation regulating the counting of electoral votes? This is a consequential question for the legal framework governing presidential elections. In 2022, Congress passed the Electoral Count Reform Act ("ECRA"), which overhauled the statutory regime governing the counting of electoral votes. The ECRA's predecessor statute, which had been in place since 1887, had long been criticized as ambiguous and unnecessarily convoluted. Those deficiencies were widely seen as a contributing cause of the attacks on the Capitol of January 6, 2021, and a rare bipartisan majority in Congress passed the ECRA to address the earlier statute's shortcomings. Yet it did so against a backdrop of unresolved questions about Congress's authority to legislate in this area. The scholarly literature, however, lacks a sustained defense of Congress's power to regulate the counting of electoral votes. This Article aims to fill that gap. It does so in two ways. First, it engages with the skeptics of Congress's authority on their traditional terrain, locating ample congressional authority grounded in the text, structure, and history of Article II and the Twelfth Amendment. The Article then seeks to expand the analytic framework by focusing on a constitutional provision that tends to stay out of the limelight: the Twentieth Amendment, which reconfigures the period between Election Day and Inauguration Day. In defending Congress's authority to pass laws regulating the counting of electoral votes, this Article provides the first scholarly treatment of the Twentieth Amendment's significance in this area.

  • Type:
    Categories:
    Sub-Categories:

    Links:

    In the current era, product differentiation is increasing; it is often fueled by big data and artificial intelligence (AI). Whereas product differentiation is generally welfare enhancing when consumers are informed and fully rational, such differentiation might reduce welfare when consumers suffer from misperceptions, either because of a lack of information or because of behavioral biases. We show that the positive and normative implications of product differentiation depend on whether consumers over or underestimate the benefits from some products. In particular, overestimation of the benefits is a potential source of significant welfare losses. We also study sellers’ incentives to promote, or combat, misperception. Our analysis can inform policymakers who are debating regulation that can make product differentiation more difficult (or easier), especially when the differentiation is instigated by AI algorithms powered by big data.