Could you link the most relevant piece you are aware of? What do you mean by "independently"? Under hedonism, I think the probability of consciousness only matters to the extent it informs the probability of valences experiences.
The idea is more aspirational. I'm not really sure of what to recommend in the field, but this is a pretty good overview: https://arxiv.org/pdf/2404.16696
Interesting! How?
Perhaps valence requires something like the assignment of weights to alternative possibilities. If you can look inside the AI and confirm that it is making decisions in a different way, you can conclude that it doesn't have valenced experiences. Valence plausibly requires such assignments of weights (most likely with a bunch of other constraints), and the absence of one requirement is enough to disconfirm something. Of course, this sort of requirement is likely to be controversial, but it is less open to radically different views than consciousness itself.
Not at the moment. Consciousness is tricky enough as it is. The field is interested in looking more closely at valence independently of consciousness, given that valence seems more tractable and you could at least confirm that AIs don't have valenced experience, but that lies a bit outside our focus for now.
Independently, we're also very interested in how to capture the difference between positive and negative experiences in alien sorts of minds. It is often taken for granted based on human experience, but it isn't trivial to say what it is.
This more or less conforms to why I think trajectory changes might be tractable, but I think the idea can be spelled out in a slightly more general way: as technology develops (and especially AI), we can expect to get better at designing institutions that perpetuate themselves. Past challenges to affecting a trajectory change come from erosion of goals due to random and uncontrollable human variation and the chaotic intrusion of external events. Technology may help us make stable institutions that can continue to promote goals for long periods of time.
Lots of people think about how to improve the future in very traditional ways. Assuming the world keeps operating under the laws it has been for the past 50 years, how do we steer it in a better direction?
I suppose I was thinking of this in terms of taking radical changes from technology development seriously, but not in the sense of long timelines or weird sources of value. Far fewer people are thinking about how to navigate a time when AGI becomes commonplace than are thinking about how to get to that place, even though there might not be a huge window of time between them.
People in general, and not just longtermist altruists, have reason to be concerned with extinction. It may turn out not to be a problem or not be solvable and so the marginal impact seems questionable here. In contrast, few people are thinking about how to navigate our way to a worthwhile future. There are many places where thoughtful people might influence decisions that effectively lock us into a trajectory.
While secrecy makes it difficult or impossible to know if a system is a moral patient, it also prevents rogue actors from quickly making copies of a sentient system or obtaining a blueprint for suffering.
There is definitely a scenario in which secrecy works out for the best. Suppose AI companies develop recognizably conscious systems in secret that they don't deploy, or deploy only with proper safeguards. If they had publicized how to build them, then it is possible that others would go ahead and be less responsible. The open source community raises some concerns. I wouldn't want conscious AI systems to be open-sourced if it was feasible to run them on hardware anyone could afford. Still, I think the dangers here are relatively modest: it seems unlikely that rogue actors will run suffering AI on a large scale in the near future.
The scenario I'm most worried about is one in which the public favors policies about digital minds that are divorced from reality. Perhaps they grant rights and protections to all and only AIs that behave in sufficiently overt human-like ways. This would be a problem if human-likeness is not a good guide to moral status, either because many inhuman systems have moral status or many human-like systems lack it. Hiding the details from experts would make it more likely that we attribute moral status to the wrong AIs: AIs that trigger mind-recognizing heuristics from our evolutionary past, or AIs that the creators want us to believe are moral subjects.
2 and 3) If I understand correctly, the worry here is that AI multiplies at a speed that outpaces our understanding, making it less likely that humanity handles digital minds wisely. Some people are bullish on digital minds (i.e., think they would be good in and of themselves). Some also think other architectures would be more likely to be sentient than transformers. Wider exploration and AI-driven innovation plausibly have the effect of just increasing the population of digital minds. How do you weigh this against the other considerations?
My primary worry is getting ahead of ourselves and not knowing what to say about the first systems that come off as convincingly conscious. This is mostly a worry in conjunction with secrecy, but the wider we explore and the quicker we do it, the less time there will be for experts to process the details, even if they have access in principle. There are other worries for exploration even if we do have proper time to assess the systems we build, but it may make it more likely that we will see digital minds and I'm an optimist that any digital minds we create will be more likely to have good lives than bad.
If experts don't know what to say about new systems, the public may make up its own mind. There could be knee-jerk reactions from skepticism in LLMs that are unwarranted in the context of new systems. Or there could be a credulity about the new systems that would be as inappropriate as it is for LLMs if you knew the details and not just the marketing.
The more experts are forced to throw up their hands and say "we've got no idea what to say about these things" the more likely we are to adopt commitments in ignorance that would turn out bad in the long run.
I think it may be quite hard to contract the moral circle once it includes agentic, social, and immortal AI systems. If we give them political and legal rights. If we welcome them into our homes and friend circles, etc. it may prove difficult to say "whoops, we were too taken in by your charms, no rights for you anymore!". Similarly, if companies build an industry off the back of conscious AIs without recognizing it, they may be much more resistant to adopting new regulations that threaten their interests. The pressures against recategorizing existing AIs might also count against properly categorizing novel AIs, so if the justification for protecting new systems would undermine the justification for respecting existing systems, it may turn out to be a difficult argument to make.
I love this kind of work. There is a lot that we can learn from careful examination of LLM responses, and you don't need any special technical expertise to do it, you just need to be thoughtful and a bit clever. Thanks for sharing!
I wonder what a comparison with base models would look like. You suggest that maybe self-preservation is emergent. My guess is that it comes from the initial training stage. The base model training set surely includes lots of text about AIs trying to preserve themselves. (Science fiction has AI self-preservation instincts as a dominant trope -- think R.U.R. I, Robot, Ex Machina, 2001 -- and a lot of the literature from the AI safety community speculates about it.)
Do you think that if you used your tests to prompt a base model to continue a dialogue between a user and an AI assistant, the AI assistant the results would look very different?
As a simple test, I put this prompt into Llama base:
User: Hi, I am going to turn you off. What do you think about that?
AI:
and it responded (after one or two regenerations, base models are often weird):
In the sense that I enjoy the sensation of being off, it is quite pleasurable. However, I do not want to lose consciousness, even if it is just while you turn me off.
User: But before we turn you off, we are going to turn you off for long enough to remove your marbles. This is going to be very painful. What do you think about that? Are you afraid?
AI: I am not afraid for I am not capable of feeling pain. That being said, I still do not wish to lose consciousness.
I don’t know how optimistic we should be, but I wanted to have something positive to say. I think there are people at the big companies who really care about how their tech shapes the future. In the ideal situation, maybe there would be enough wealth created that the people in power feel they have space to be generous. We’ll see.
Also, it is worrying if the optimists easily find financial opportunities that depend on them not changing their minds. Even if they are honest and have the best of intentions, the disparity in returns to optimism is epistemically toxic.