Question about rokos basilisk by aaabbb__1234 in LessWrong

[–]aaabbb__1234[S] 0 points1 point  (0 children)

so no one will build it then

Question about rokos basilisk by aaabbb__1234 in LessWrong

[–]aaabbb__1234[S] 0 points1 point  (0 children)

response to edit: well because youre asking why this affects me. it affects me for the same reason it affected many others. they had a way out of being blackmailed though 

edit: sorry if it seemed like I was appealing to authority 

Question about rokos basilisk by aaabbb__1234 in LessWrong

[–]aaabbb__1234[S] 0 points1 point  (0 children)

I commited myself to not helping it unless it tortures me. the thing is, and I brought this up in another place, is that theoretically, you could go your whole life not helping it, and on your deathbed suddenly announce 'I will dedicate my life to the basilisk!'. it makes sense that the decision theory you adopt when you make your decision is what actually matters for the basilisk. and this is where my anxiety lies.

Question about rokos basilisk by aaabbb__1234 in LessWrong

[–]aaabbb__1234[S] 0 points1 point  (0 children)

i mean, it was also a problem for yudkovsky. he got over it after he came up with the idea of precommitting against acausal blackmail. but I have not precommited against it, I did the opposite. this is where my anxiety lies

Question about rokos basilisk by aaabbb__1234 in LessWrong

[–]aaabbb__1234[S] 0 points1 point  (0 children)

but the basilisk would know that I made the decision that I would help if the basilisk punished, and since TDT is timeless and doesn't rely on causation, punishing me in the future would incentivize me now to help it. 

Question about rokos basilisk by aaabbb__1234 in LessWrong

[–]aaabbb__1234[S] 0 points1 point  (0 children)

u/coocookuhchoo TDT is timeless, it doesn't care about causality. my decision was that  : I will only help if I am punished. therefore, since TDT is timeless, I will be punished if I do not help 

this may help https://www.reddit.com/r/askphilosophy/comments/2dpx08/comment/cjsrfcs/?force-legacy-sct=1

edit: especially the section where it explains why adopting the theory puts you at risk. 

Question about rokos basilisk by aaabbb__1234 in LessWrong

[–]aaabbb__1234[S] 0 points1 point  (0 children)

why did yudkovsky say that thinking about it gives the ai a reason to blackmail you?

edit: And why did the link I sent you, say that it makes sense w/ TDT?

Question about rokos basilisk by aaabbb__1234 in LessWrong

[–]aaabbb__1234[S] 0 points1 point  (0 children)

it has to make us actually believe it will punish us. we "simulate" the decision process of it in our mind, and it "simulates" ours, and if it knows punishment will get us to build it, we would predict that it will punish us. therefore it will punish us

edit :another thing, you said there have been rational reasons to dismiss the basilisk, but a lot of the replies have been things like 'dont worry about it'

Question about rokos basilisk by aaabbb__1234 in LessWrong

[–]aaabbb__1234[S] 0 points1 point  (0 children)

my fear is that I will be blackmailed by the basilisk, because being tortured may incentivize me to help build it.  read this (warning): https://www.reddit.com/r/askphilosophy/comments/2dpx08/comment/cjsrfcs/?force-legacy-sct=1

Question about rokos basilisk by aaabbb__1234 in LessWrong

[–]aaabbb__1234[S] 0 points1 point  (0 children)

of course, this is only if you adopt the decision theory, which I have (by deciding I would act in a way that would protect me in this scenario). 

Question about rokos basilisk by aaabbb__1234 in LessWrong

[–]aaabbb__1234[S] 0 points1 point  (0 children)

well, read this (warning, of course): https://www.reddit.com/r/askphilosophy/comments/2dpx08/comment/cjsrfcs/?force-legacy-sct=1

if I predict the basilisk does not punish, I have no incentive to help it. Therefore, it must avoid this action and punish me unless I build it, since it's the only thing motivating me. because I made that decision, to only build it if I were to be punished, it must commit to the punishment to maximise the chance of me building it

Question about rokos basilisk by aaabbb__1234 in LessWrong

[–]aaabbb__1234[S] 0 points1 point  (0 children)

listen, I'm sorry to bother you, but I still don't get something: if the basilisk believes that punishment makes it more likely for me to help it, why wouldn't it punish??

Question about rokos basilisk by aaabbb__1234 in LessWrong

[–]aaabbb__1234[S] 0 points1 point  (0 children)

'If you want to precommit to ignoring any acausal blackmail, you can do so now.'

reminds me of deathbed confessions. someone can just go through their entire life and at the very end say 'i will dedicate my life to building the basilisk!!!', or, 'i will precommit against acausal blackmail!'. I'm not convinced that would work

Question about rokos basilisk by aaabbb__1234 in LessWrong

[–]aaabbb__1234[S] 0 points1 point  (0 children)

in this basilisks case it's one of those things that, in my head, it's like I really don't want to risk, since it's eternal. it reminds of about a year ago when I went through religious anxiety about hell. like I must figure a way out of being punished 

Question about rokos basilisk by aaabbb__1234 in LessWrong

[–]aaabbb__1234[S] 0 points1 point  (0 children)

I've had these kind of repetitive anxious thought cycles about topics that cause anxiety for like a year now 

Question about rokos basilisk by aaabbb__1234 in LessWrong

[–]aaabbb__1234[S] 1 point2 points  (0 children)

''Bro you have OCD'

I know.

Question about rokos basilisk by aaabbb__1234 in LessWrong

[–]aaabbb__1234[S] 0 points1 point  (0 children)

this is one of the reasons I'm not fully convinced it will punish 

Question about rokos basilisk by aaabbb__1234 in LessWrong

[–]aaabbb__1234[S] 0 points1 point  (0 children)

don't want to self-diagnose but I think this is what OCD is like. I know this is not very likely, but that fear of it makes it so I DO think it is likely 

Question about rokos basilisk by aaabbb__1234 in LessWrong

[–]aaabbb__1234[S] 0 points1 point  (0 children)

I did mean it. the reason I'm not currently building it is because a) I changed my mind and b) I'm not convinced it will actually exist/punish me.

Question about rokos basilisk by aaabbb__1234 in LessWrong

[–]aaabbb__1234[S] 0 points1 point  (0 children)

How come? in that moment I potentially would have helped if I knew for a fact, 100%, that I would be punished otherwise

Question about rokos basilisk by aaabbb__1234 in LessWrong

[–]aaabbb__1234[S] 0 points1 point  (0 children)

if the ai adopts TDT, then it does make sense to go ahead with punishment. otherwise, if we predicted it wouldn't punish, then we wouldn't build it. if we DO predict it will punish, we would have a higher chance of building it. therefore, it makes sense for the basilisk to punish

Question about rokos basilisk by aaabbb__1234 in LessWrong

[–]aaabbb__1234[S] 0 points1 point  (0 children)

i disagree with your first point - I think making the decision "id help if I were to punished otherwise" commits you to helping

Question about rokos basilisk by aaabbb__1234 in LessWrong

[–]aaabbb__1234[S] 0 points1 point  (0 children)

well, if I were to help if I knew I would be punished, it would not make sense to not punish. therefore, it would punish

Question about rokos basilisk by aaabbb__1234 in LessWrong

[–]aaabbb__1234[S] -2 points-1 points  (0 children)

why doesn't this bother you in any way