07 January 2023

powerful quotes I keep contemplating

last updated: 7 January 2023


The question is not, Can they reason? nor, Can they talk? but, Can they suffer?

~ Jeremy Bentham (in An Introduction to the Principles of Morals and Legislation)


The smart way to keep people passive and obedient is to strictly limit the spectrum of acceptable opinion, but allow very lively debate within that spectrum—even encourage the more critical and dissident views. That gives people the sense that there's free thinking going on, while all the time the presuppositions of the system are being reinforced by the limits put on the range of the debate.

~ Noam Chomsky (in The Common Good)


The total amount of suffering per year in the natural world is beyond all decent contemplation. During the minute that it takes me to compose this sentence, thousands of animals are being eaten alive, many others are running for their lives, whimpering with fear, others are slowly being devoured from within by rasping parasites, thousands of all kinds are dying of starvation, thirst, and disease. It must be so. If there ever is a time of plenty, this very fact will automatically lead to an increase in the population until the natural state of starvation and misery is restored. In a universe of electrons and selfish genes, blind physical forces and genetic replication, some people are going to get hurt, other people are going to get lucky, and you won't find any rhyme or reason in it, nor any justice. The universe that we observe has precisely the properties we should expect if there is, at bottom, no design, no purpose, no evil, no good, nothing but pitiless indifference.

~ Richard Dawkins (in River Out of Eden: A Darwinian View of Life)


Custom will reconcile people to any atrocity; and fashion will drive them to acquire any custom.

~ George Bernard Shaw (in Killing For Sport)


The real problem of humanity is the following: we have paleolithic emotions, medieval institutions, and god-like technology.

~ Edward Osborne Wilson (in a debate at the Harvard Museum of Natural History)


There are no catastrophes that loom before us which cannot be avoided; there is nothing that threatens us with imminent destruction in such a fashion that we are helpless to do something about it. If we behave rationally and humanely; if we concentrate coolly on the problems that face all of humanity, rather than emotionally on such nineteenth century matters as national security and local pride; if we recognize that it is not one’s neighbors who are the enemy, but misery, ignorance, and the cold indifference of natural law—then we can solve all the problems that face us. We can deliberately choose to have no catastrophes at all.

~ Isaac Asimov (in A Choice of Catastrophes: The Disasters that Threaten Our World)


Disaster on an unfathomable scale is always taking place on Earth. Countless instances of extreme suffering are occurring in this moment — *right now*. Yet because this suffering is so normal and ordinary, simply occurring every day, distributed rather evenly over time and space, it seems less evocative and urgent than the more unusual, more localized disasters, such as school shootings and earthquakes. Almost all the suffering that occurs on Earth can be considered baseline horror, which allows us to ignore it. We simply do not *feel* the ever-present emergency that surrounds us.

~ Magnus Vinding (in Suffering-Focus Ethics: Defense and Implications)


It's easy to convince oneself that things can't *really* be that terrible, that the horror I allude to is being overblown, that what is going on elsewhere in space-time is somehow less real than the here-and-now, or that the good in the world somehow offsets the bad. Yet however vividly one thinks one can imagine what agony, torture or suicidal despair must be like, the reality is inconceivably worse. The force of "inconceivably" is itself largely inconceivable here. Blurry images of Orwell's "Room 101" can barely even hint at what I'm talking about.

~ David Pearce (in The Hedonistic Imperative)


We advocate the well-being of all sentience, including humans, non-human animals, and any future artificial intellects, modified life forms, or other intelligences to which technological and scientific advance may give rise.

~ Humanity+ (in The Transhumanist Declaration)


Health is a state of complete physical, mental and social well-being and not merely the absence of disease or infirmity.

~ in Constitution of the World Health Organization


Had Mother Nature been a real parent, she would have been in jail for child abuse and murder.

~ Nick Bostrom (in In Defense of Posthuman Dignity)


To be hopeful in bad times is not just foolishly romantic. It is based on the fact that human history is a history not only of cruelty, but also of compassion, sacrifice, courage, kindness.
What we choose to emphasize in this complex history will determine our lives. If we see only the worst, it destroys our capacity to do something. If we remember those times and places–and there are so many–where people have behaved magnificently, this gives us the energy to act, and at least the possibility of sending this spinning top of a world in a different direction.
And if we do act, in however small a way, we don’t have to wait for some grand utopian future. The future is an infinite succession of presents, and to live now as we think human beings should live, in defiance of all that is bad around us, is itself a marvelous victory.

~ Howard Zinn (in You Can’t Be Neutral on a Moving Train

06 January 2023

Moorean Metaethics and Evolutionary Metaethics: Moore’s Principia Ethica facing Street’s Darwinian Dilemma (term paper)

During my fourth semester of studying philosophy at the University of Zurich, I attended a seminar titled "G.E. Moore: Some Main Problems of Philosophy", which was run by Peter Schultewho introduced us to the eponymous book. I was intrigued by Moore's work on the role of intuition and common sense in philosophy, which further fuelled my interest in epistemology, especially moral epistemology. As I had also been meaning to explore evolutionary debunking arguments for quite some time, I decided to write a term paper on Moorean metaethics and evolutionary metaethics during the subsequent semester.

This is the link to my second term paper in philosophy: Moorean Metaethics and Evolutionary Metaethics: Moore’s Principia Ethica facing Street’s Darwinian Dilemma

20 July 2022

Surveying Pure Time Preference among the Swiss Electorate for Intergenerational Social Cost Benefit Analysis (research proposal)

During my fourth semester of studying political science at the University of Zurich, I attended a Public Choice seminar titled "Risk Preferences and Economic Development", which was run by Philipp Kerler. One of the many topics we touched upon was time discounting, which I was immediately interested in because of its importance in the context of longtermism, so I picked time discounting as topic for the research proposal we had to write.

This is the link to my first research proposal in political science: Surveying Pure Time Preference among the Swiss Electorate for Intergenerational Social Cost Benefit Analysis


04 January 2022

philosophy event series introducing EA at the University of Zurich

< < < see this post on the EA Forum here > > >

I was given the opportunity to organise and run a student-led university course introducing attendees to the philosophy and social movement of EA. :)

Even though it was officially categorised as a philosophy course, I would describe it as "an introductory reading and discussion group with a focus on philosophy".

Prior to running the course, I had sought feedback on a preliminary collection of resources for the syllabus and, having incorporated three people's feedback, now think this should give a more well-rounded, balanced intro into mainstream EA. Prior to that, my preliminary collection of resources for the syllabus was too suffering-focused for it to be a well-rounded, balanced introduction into what mainstream EA currently is.

The event series took place at UZH (University of Zurich) in the last autumn semester.
For six events, a guest with more background knowledge than me in that day's topic was invited by me and arranged to make it to the event to enrich the discussion.
There were varying but low numbers of attendees, with the most being four and with only one attendee attending the whole event series. If I ever rerun such an event series, I won't schedule it for Friday from 16:15 to 17:45 and I'll start advertising it sooner.

I think it can't hurt to share the syllabus below and my power point slides with discussion prompts here.

Maybe it's useful for someone who has a similar opportunity. Or for people looking into a philosophy-focused intro to EA. And if I ever decide to rerun such an event series, I'd make sure to take any suggestions / criticisms / ... of yours to heart prior to doing so. :)

These were my syllabus-related preliminary remarks:

I distinguish between resources designated as
E1: expected
E2: encouraged
E3: extra
Additionally, I would like to encourage you to quickly scan the resources you are not going to read in full, especially those I encourage you to read (i.e. “E2”).
I recommend reading/watching the resources in the order I put them.

Table of Contents

session #1: Introduction

session #2: Global Health and Development

session #3: Animal Welfare

session #4: The Long-Term Future

session #5: Putting EA into Practice – Part 1/2

session #6: Global Catastrophic Biological Risks

session #7: Positively Shaping the Development of Artificial Intelligence

session #8: Global Priorities Research and Moral Uncertainty

session #9: Putting EA into Practice – Part 2/2

session #1: Introduction

session #2: Global Health and Development

session #3: Animal Welfare

session #4: The Long-Term Future

session #5: Putting EA into Practice – Part 1/2

session #6: Global Catastrophic Biological Risks

session #7: Positively Shaping the Development of Artificial Intelligence

session #8: Global Priorities Research and Moral Uncertainty

session #9: Putting EA into Practice – Part 2/2

15 September 2021

the metaethics and normative ethics of AGI value alignment: many questions, some implications (commentated research agenda)

< < < see this post on the EA Forum here > > >

Short Summary

This (<--) is a commentated research agenda on the metaethics and normative ethics of AGI (artificial general intelligence) value alignment. It aims to show that various crucial aspects of this interdisciplinary Herculean task hinge on

  1. numerous metaethical (and metaethics-related) questions;
  2. on the moral implications of different answers to these questions;
  3. on how relevant actors end up deciding to take the uncertainty surrounding these (meta)ethical questions into account.

My key takeaways (for the short summary):

  1. At least some of the questions about metaethics and normative ethics briefly discussed in the six main parts and listed in the two appendices merit further attention.
  2. Let us take seriously Nick Bostrom’s verdict that we have to do philosophy with a deadline.
  3. Let us get a better, more exhaustive understanding of what work on AGI value alignment currently does involve, and what it could involve, and what it should involve, as well as the most decision-relevant uncertainties, and how to mitigate these uncertainties, and how to handle the remainder.
  4. Let us give not only the intradisciplinary work but also the interdisciplinary work that the challenge of AGI value alignment requires its due attention.

Longer Summary

(I assume reading the short summary preceded starting to read this sentence.)

Artificial general intelligence (AGI) is a form of artificial intelligence exhibiting the capacity to apply its problem-solving skills in a broadly human-comparable cross-domain fashion (as opposed to its intelligence being forever narrowly restricted to a given set of tasks).

Since AGI, although currently hypothetical, is thought of by most pertinent researchers as “eventually realised” instead of “forever mere sci-fi”, there is ample need to research and discuss the implications this emerging technology might have. One of the most important questions arising in this context is how to go about AGI value alignment. This is among the most prominent interventions for increasing AGI safety and involves researching how to make AGI(s) “aligned to humans by sharing our values”. Incidentally: What exactly this even means gives rise to foundational questions that are far from settled.

There is a beautiful Value Alignment Research Landscape map by the Future of Life Institute giving an impressive overview of the interdisciplinary nature of this Herculean task. In addition to the technical aspects of AGI value alignment, there are also socio-politico-economic as well as neuro-bio-psychological, and philosophical aspects.

Of all the literature on questions in moral philosophy that artificial intelligence gives rise to, only a small share is concerned with the intersection of metaethics and artificial general intelligence. Metaethics is the discipline concerned with higher-order questions about ethics such as what the nature of morality is, what moral goodness/badness means, whether and if so, to which extent and in which form such a thing even exists, and how we should “do ethics”. At various places throughout this text, I employ the term metaethics in a very broad sense, e.g. as overlapping with moral psychology or the philosophy of mind.

In addition to two appendices (a list of further research questions and a list of lists of further research questions), there are six main parts:

1) AGIs as Potentially Value Aligned

I briefly touched upon these questions:

I think these questions are very important, and/but do not think any of this should make us completely abandon any hope of achieving anything in the way of AGI value alignment.

2) AGIs as Potential Moral Agents and as Potential Moral Patients

  • Which conditions have to be met for an artificial entity to be(come) a moral agent, i.e. an entity that, due to being capable of moral deliberation, has moral reasons to behave in some ways but not in others?
  • Which conditions have to be met for an artificial entity to be(come) a moral patient, i.e. an entity that should be behaved towards (by moral agents) in some ways but not in others?

As of now, we know of no entity that is widely agreed upon to be a moral agent but is widely agreed upon not to be a moral patient. The prospect of the eventual advent of some forms of artificial minds might change that. It is possible, though, that AGI will not change anything about the emptiness of this fourth category: Firstly, moral deliberation seems necessary for moral agency. Secondly, it seems possible that sentience might be necessary for (robust forms of) moral deliberation. Thus, moral agency (at least in some substantive manifestation) might perhaps be impossible without an affective first-person understanding of sentience, i.e. without experiencing sentience.

We can add to that a third premise: It seems possible that artificial sentience is not going to happen. It is thus not obvious whether any AGI will ever be(come) a moral agent in anything but at most a minimalist sense. However, none of this should make us confident that artificial moral agency (with or without moral patiency) is impossible (or possible only in some minimalist sense of moral agency) either.

More clarity on what it is that humans are doing when we are engaging in moral deliberation seems to be useful for determining what it would mean for an artificial entity to engage in moral deliberation. Closely related, getting a better understanding of necessary conditions
and of sufficient conditions for moral agency and moral patiency seems to be useful for getting a better understanding of which actual and which hypothetical artificial entities (might come to) qualify as moral agents and/or moral patients.

While the philosophy of mind is not usually thought of as being within the scope of moral philosophy, I would argue there is significant overlap between the philosophy of mind and some metaethical questions. The prospect of the eventual advent of some forms of artificial minds raises questions such as whether moral agents can exist without being moral patients and/or without being sentient and/or without being sapient. Satisfyingly answering metaethical questions related to the chances and risks of creating entities that are more ethical than its creators might thus also presuppose a better understanding of consciousness.

3) AGIs as Potentially More Ethical Than its Creators

Here, I covered three topics overlapping with metaethics that are relevant to creating AGI more ethical than its creators: The orthogonality thesis, moral truth(s), and value lock-in.

3a) The Orthogonality Thesis

The orthogonality thesis denies that we should expect higher intelligence to automatically go hand in hand with “higher/better goals”. While even some AGI experts are at least sceptical of the orthogonality thesis or the decision relevance of its implications, this is one of the few things in this text that I have an opinion about. The position I argued for is that rejecting the orthogonality thesis would be so (net-)risky that we should only do so if we were extremely confident in a universal and sufficiently strong form of internalism re moral motivation, but that we should not have this confidence. We should not assume that whatever AGI we create would feel compelled (without our having done anything to make it so) to act in accordance with the moral truths (it thinks) it is approaching. Thus, I stated that the challenge of creating AGI more ethical than its creators does not just encompass it be(com)ing better than they or we humans generally are at moral deliberation but at altogether “being good at being good”.

3b) Moral Truth(s)

I briefly explained three forms of moral scepticism but went on to argue that, to the extent that humanity is committed to arriving at or at least approximating moral truth(s), there seems to be no such thing as abandoning this mighty quest too late.

I then brought up a “family” of metaethical theories that I deem to be particularly interesting in the context of AGI value alignment: metaethical theories, universalist as well as relativist, involving idealisation (of, e.g. values).

3c) Value Lock-in

I understand a value lock-in to be a scenario where some process has rendered future changes of prevailing values of agents exerting considerable world-shaping behaviour qualitatively less feasible than before, if at all possible. There are AGI-related risks of deliberately or inadvertently locking in values that later turn out to be “wrong values” – if moral realism or moral constructivism is true and we arrive at such truth(s) – or simply values we come to wish (for whatever reason) we had not locked in. Thus, we should think that the creator(s) of the first AGI setting into stone their values (or, for that matter, some aggregation of the values of humanity at that time) is not the best course of action…

4) Unaligned AI, Self-Defeating Cosmopolitanism, and Good AI Successors

… Nor should we, on the other extreme, pursue a deliberate AGI “non-alignment”, which, given the likelihood of a sufficiently intelligent such AGI taking over and not being cosmopolitan at all, embodies what I termed self-defeating cosmopolitanism.

Questions touched upon in this part:

5) Lovecraftian Ethics

Considerations about the far future of Earth-originating intelligence, and Earth-originating moral patients and moral agents prompt numerous mighty questions varyingly evocative of Lovecraftian horror. It is in the nature of things that, as we continue to study consciousness and the cosmos, making scientific discoveries and exploring new philosophical questions, we will probe more and more of what I termed Lovecraftian ethics. Two examples:

What is unclear is not just what questions and answers we will encounter, but also how, if at all, our current primitive Darwinian minds can and will cope with what as of yet lies at the border of the conceivable or wholly beyond. What can already be said now, however, is this: The future of intelligence is a major determining factor of both the future of “doing ethics” and the future of (dis)value in the universe. And we can influence the future of intelligence for the better or die trying.

6) Dealing with Uncertainties

I speculated whether aiming for a universal theory of uncertainty might be a worthwhile pursuit and pointed out that there might very well be, due in part to path dependencies, an urgency in finding better answers to philosophical questions, and in turn listed some more questions that we should probably try to tackle sooner rather than later.

Given the empirical uncertainty as well as the uncertainty re normative ethics and metaethics (and meta levels above that), we would ideally lock in as few things as possible (to preserve option value) but as many as needed (to avoid catastrophes). There are at least two extremes we should avoid: earliest possible value lock-in and deliberate “non-alignment”. The most promising family of approaches for AGI value alignment seems to me to centre around ideas such as uncertainty, indirectness, plasticity, and idealisation of values / interests / desires. Broadly speaking, the best course of action appears to involve

  • starting with the best we have to offer – instead of insisting we know too little about ethics and are too immoral to do that
  • and extrapolating from that in the best possible way we can think of – instead of insisting that we need not do this since we already know enough about ethics and are already moral enough.

Additionally, I think it is crucial that more work is being done on uncertainty re normative ethics and metaethics (and meta levels above that). I consider it extremely unlikely that anyone will be morally justified in being uncompromising when it comes to solving the value definition problem, i.e. deciding which values the first AGI will have (assuming the value loading problem, the technical problem of how to do this, has been fully solved by then – and assuming, probably unjustifiably, that these are fully separate problems). There seems not to be enough time left to solve normative ethics – if this is possible without AGI (or maybe human enhancement) in the first place. To tie in with this: That current approaches to value alignment seem to be more readily compatible with consequentialism than with deontological ethics or virtue ethics seems to me like a comparably concrete and tractable problem.

I want to reiterate that this text is not an attempt at arguing at length for a definite answer to any of these questions. Rather, I imagine this text to serve as a commentated research agenda explaining its own purpose and hopefully inspiring some more people to try to tackle some of these mighty questions.

Acknowledgements

This text is my deliverable of the first summer research program by the Swiss Existential Risk Initiative (CHERI). I am very grateful to the organisers of CHERI for their support. Special thanks to my CHERI-assigned mentor, François Jaquet, who helped me navigate this project. I started with but a few dozens of hours of background knowledge in metaethics (the same applies, for that matter, to AGI), which is partly why our weekly video calls and his feedbacks proved really valuable. Views and mistakes are my own.

19 August 2021

suffering-focused population ethics (term paper)

During my first semester of studying philosophy at the University of Zurich, I attended a seminar titled "Ethics of the Future", which was run by Stefan Riedener. He introduced us to many fascinating and important topics, including Derek Parfit's pioneering work on population ethics. I decided to write a term paper on the Repugnant Conclusion from a suffering-focused perspective during the subsequent semester.

Please note that considerations re moral uncertainty and cooperation (among other things) keep me from embracing courses of action that the vast majority of my contemporaries deem extremely immoral.

If it had not been for David Pearce's comment that "it deserves to be a public resource on the Net so it can be linked/cited", I would be (even) more hesitant on sharing this publicly, but here go you.

powerful quotes I keep contemplating

last updated: 7 January 2023 The question is not, Can they reason? nor, Can they talk? but, Can they suffer? ~ Jeremy Bentham (in An Introdu...