KD

Kenneth_Diao

244 karmaJoined

Comments
17

I'm not an expert, but this may be a good idea. Apparently ranked-choice voting is always vulnerable to certain types of failures (https://en.wikipedia.org/wiki/Arrow%27s_impossibility_theorem), but these can be avoided with rated voting systems.

The donation election post (meet the candidates) and the actual voting platform need to be cross-checked. I saw that Animetrics was included in the vote but not in the post, while Giving Green was included in the post and not in the vote. There may be other errors which I missed.

I voted for mainly animal welfare/rights charities first, particularly ones which focused on highly neglected, large-scale populations like insects, shrimps, and fishes. I also voted highly for PauseAI because I believe in creating greater public pressure to slow AI progress and shifting the Overton Window, even if I am agnostic about pausing AI progress itself. After these, I voted for some of the meta/mixed organizations which I thought were especially promising, including Rethink Priorities and the Unjournal. Then I voted for mental health/resilience interventions. Then I voted for GCR initiatives. I did not vote for any human welfare interventions which I expected to cause net harm to animals. I did not vote for any other AI organizations because I did not trust that they were sufficiently decelerationist.

 I I think this is an interesting dilemma, and I am sympathetic to some extent (even as an animal rights activist). At the heart of your concern are 3 things:

  1. Being too radical risks losing popular support
  2. Being too radical risks being wrong and causing more harm than good
  3. How do we decide what ethical system is right or preferable without resorting to power or arbitrariness?

I think in this case, 2) is of lesser concern. It does seem like adults tend to give far more weight to humans than animals (a majority of a sample would save 1 human over 100 dogs), though interestingly children seem to be much less speciesist (Wilks et al., 2020). But I think we have good reasons to give substantial moral weight to animals. Given that animals have central nervous systems and nociceptors like we do, and given that we evolved from a long lineage of animals, we should assume that we inherited our ability to suffer from our evolutionary ancestors rather than uniquely developing it ourselves. Then there's evidence, such as (if I remember correctly) that animals will trade off material benefits for analgesics. And I believe the scientific consensus has consistently and overwhelmingly been that animals feel pain. Animals are also in the present and the harms are concrete, so animal rights is not beset by some of the concerns that, say, long-termist causes are. So I think the probability that we will be wrong about animal rights is negligible.

I sympathize with the idea that being too radical risks losing support. I've definitely had that feeling myself in the past when I saw animal rights activists who preferred harder tactics, and I still have my disagreements with some of their tactics and ideas. But I've come to see the value in taking a bolder stance as well. From my experience (yes, on a college campus, but still), many people are surprisingly willing to engage with discussions about animal rights and about personally going vegan. Some are even thankful or later go on to join us in our efforts to advocate for animals. I think for many, it's a matter of educating them about factory farming, confronting them with the urgency of the problem, and giving them space to reflect on their values. And even if you don't believe in the most extreme tactics, I think it's hard to defend not advocating for animal rights at all. Just a few centuries ago, slavery was still widely accepted and practiced, and abolitionism was a minority opinion which often received derision and even threats of harm. The work of abolitionists was nevertheless instrumental in getting society to change its attitudes and its ways such that the average person today (at least in the West) would find slavery abhorrent. Indeed, people would roundly agree that slavery is wrong even if they were told to imagine that the enslaved person's welfare increased due to their slavery (based on a philosophy class I took years ago). To make progress toward the good, society needs people who will go against the current majority.

And this may lead to the final question of how we decide what is right and what is wrong. This I have no rigorous answer to. We are trapped between the Scylla of dogmatism and the Charybdis of relativism. Here I can only echo the point I made above. I agree that we must give some weight to the majority morality, and that to immediately jump ten steps ahead of where we are is impractical and perhaps dangerous. But to veer too far into ossification and blind traditionalism is perhaps equally dangerous. I believe we must continue the movement and the process towards greater morality as best we can, because we see how atrocious the morality of the past has been and the evidence that the morality of the present is still far from acceptable.

I know this is a debate, but one thing I want to touch on is that animal welfare and human welfare are not necessarily in conflict. I think initiatives like preventing the rise of factory farming in the developing world could be really great for both animals and humans. Animals wouldn't have to exist in horrible conditions, and humans could (as far as I know; don't have sources with me right now) have greater food, water, and resource security, reduced ecological/climate devastation, and reduced risk of disease, to name a few things. I think it's important to think about ways in which we can jointly improve animal welfare and global health, because we all ultimately want to create a better world.

A few reasons immediately come to mind for me:

  1. There are many more animals in factory farms than humans (scale)
  2. The average suffering of these animals is likely worse than the average suffering of humans (because animals are almost uniformly kept in horrendous conditions, while humans are not) (scale)
    1. My intuition is that the "moral multiplier" of human ability to suffer is not much higher than 1, if at all, for many animals. Animals have central nervous systems and nociceptors just like we do. Mammal suffering in particular might be close to par with humans, but I see no obvious reason that birds or fish are somehow less able to suffer. I also think that there's probably some bias due to our culture's philosophical heritage of "rational capability = moral consideration"
    2. Not an expert at this, though, so it's just me freewheeling
  3. I don't have exact numbers with me, but I would bet that animal welfare/rights receives much less funding and attention than global health and development (neglectedness)
    1. I've also heard that a dollar could prevent more years of, say, chicken suffering than years of human suffering (tractability)

For me, I think the biggest crux is whether you believe animal suffering is comparable to human suffering. Animal is a broad category, but I think at least for some animals, there is all the reason to think that their suffering is comparable and little reason to think it is not. The only reason I put one notch below the maximum is to signal that I am willing to concede some slight uncertainty about this, but nowhere near enough to persuade me that animal welfare/rights is not a pressing cause.

Thanks Gabe! Yes, I agree that aligning to the right values in the right way which will be widely accepted as legitimate is a pretty deep and broad problem.

Hi Leopold,

Thank you for the thoughtful comment! I appreciate that my experience has informed your decision-making, but in the end it’s just my experience, so take it with a grain of salt. I also appreciate your caution; I would say that I’m also a pretty cautious person (especially for an EA; I personally think we sometimes need a little more of that).

I will say that big and risky projects aren’t necessarily a bad thing; they’re just big and risky. So if you’ve carefully considered the risks and acknowledged that you’re committing to a big project that might not pay off and you have some contingency plans, then I think it’s fine to do. I just think that sometimes we get caught up in the vision and end up goodharting for bigger and more visionary projects rather than more actually effective ones (my failure mode in Spring 2023).

Best, Kenneth

This kind of reminds me of a psychological construct called the Militant Extremist Mindset. Roughly, the mindset is composed of three loosely related factors: proviolence, vile world, and Utopianism. The idea is that elevated levels in each of the three factors is most predictive of fanaticism I think (total) utilitarianism/strong moral realism/lack of uncertainty/visions of hedonium-filled futures fall into the utopian category. I think EA is pretty pervaded but vile world thinking, including reminders about how bad the world is/could be and cynicism about human nature. Perhaps what holds most EAs back at this point is a lack of proviolence—a lack of willingness to use violent means/cause great harm to others; I think this can be roughly summed up as “not being highly callous/malevolent”.

I think it’s important to reduce extremes of Utopianism and vile world in EA, which I feel are concerningly abundant here. Perhaps it is impossble/undesirable to completely eliminate them. But what might be most important is something that seems fairly obvious: try to screen out people who are capable of willfully causing massive harm (i.e., callous/malevolent individuals).

Based on some research I’ve done, the distribution of malevolence is relatively highly right-skewed, so screening for malevolence probably affects the fewest individuals while still being highly effective. It also seems that callousness and a willingness to harm others for instrumental gain are associated with abnormalities in more primal regions of the brain (like the Amygdala) and are highly resistant to interventions. Therefore, changing the culture is very unlikely to robustly “align” them. And intuitively, a willingness to cause harm seems to be the most crucial component, while the other components seem to be more channeling malevolence towards a more fanatical bent.

Sorry I’m kind of just rambling and hoping something useful comes out of this.

Load more