I'm not sure that really responds to my point.
You seem to think that we should only posit qualia if it's strictly necessary to observe a behaviour. I was pushing back by saying that, given evolutionary relatedness and observed similarities, I don't believe that should be the criteria.
For example, you could try to explain a baby's pain behaviour without reference to qualia as not being strictly necessary, but given that we know that they are just developing versions of us, that's not the most likely explanation. With other animals obviously the inference there is weaker, but still that's a factor that should be considered.
As for specific problems. You can make plausible arguments for lots of different features. I don't think it's too useful to hyperfocus on specific ones. Protective behaviour (limping/wound garden) is one candidate that the Rethink Priorities reports discuss that I think is stronger evidence, since it's non-reflexive, long-term and involves motivational trade-offs.
The question is not whether these behaviours could strictly be explainable without qualia. The question as what's the most likely explanation given that these animals are related to us and we solve a lot of these problems through qualia (while showing similar external signs).
For example, yeah, a dog could just look like she is in pain. But then we have to invent this new concept of ersatz pain that looks and functions a lot like our pain, but is actually unconscious, in order to describe the dogs mental state in this case. To the extent it looks similar to our pain in a given animal, this looks like a ad hoc move.
Seems to me like the case for the importance of verbal report as evidence may be overstated. If my friend looks and behaves like they are suffering (cringing, recoiling away, etc.), but says they are not, I would assume that they actually are suffering. I assume you would agree? That suggests that other forms of evidence are actually more important than talking about qualia.
I don't think the scale or expected value affects this strategy question directly. You still just use a strategy that is going to be most likely to achieve the goal.
If the goal is something you have really widespread agreement on, that probably leans you towards an uncompromising, radical ask approach. Seems like things might be going pretty well for AI safety in that respect, though I don't know if it's been established that people are buying into the high probability of doom arguments that much. I suspect that we are much less far along than the climate change movement in that respect, for example. And even if support were much greater, I wouldn't agree with a lot of this post.
Oh, my expertise is in animal advocacy, not AI safety FYI
I think this is tantamount to saying that we shouldn't engage within the political system, compromise, or meet people where they are coming from in our advocacy. I don't think other social movements would have got anywhere with this kind of attitude, and this seems especially tricky with something very detail orientated like a AI safety.
Inside game approaches (versus outside game approaches like this is describing) are going to require engaging in things this post says that no one should do. Boldly stating exactly the ideal situation you are after could have its role, but I'd need to see and much more detailed argument about why that should be the only game in town when it comes to AI.
I think that as AI safety turns more into an advocacy project it needs engage more with the existing literature on the subject including what has worked for past social movements.
Also, importantly, this isn't lying (as Daniel's comment explains).
Unfortunately, I'm not aware of something like that.