I can’t believe we’re doing this again.
-
I can’t believe we’re doing this again. It’s just a bot that generates the text you ask it for. If you put it in charge of critical decisions, it will kill people. Not because it’s secretly evil, but because it’s a word generator. It’s like putting your toaster in charge of air traffic control.
-
I can’t believe we’re doing this again. It’s just a bot that generates the text you ask it for. If you put it in charge of critical decisions, it will kill people. Not because it’s secretly evil, but because it’s a word generator. It’s like putting your toaster in charge of air traffic control.
@malwaretech in the the old days the toasters were flying to save the screens.

-
I can’t believe we’re doing this again. It’s just a bot that generates the text you ask it for. If you put it in charge of critical decisions, it will kill people. Not because it’s secretly evil, but because it’s a word generator. It’s like putting your toaster in charge of air traffic control.
@malwaretech News flash: “software trained to reproduce text similar to text that commonly occurs in a given situation, including lots of fictional AI dystopias, reproduces text that is similar to what commonly occurs in discussions of possible AI dystopias”
-
@malwaretech in the the old days the toasters were flying to save the screens.

-
I can’t believe we’re doing this again. It’s just a bot that generates the text you ask it for. If you put it in charge of critical decisions, it will kill people. Not because it’s secretly evil, but because it’s a word generator. It’s like putting your toaster in charge of air traffic control.
@malwaretech@infosec.exchange That makes sense.
-
I can’t believe we’re doing this again. It’s just a bot that generates the text you ask it for. If you put it in charge of critical decisions, it will kill people. Not because it’s secretly evil, but because it’s a word generator. It’s like putting your toaster in charge of air traffic control.
@malwaretech@infosec.exchange "so called expert prompts a text gen bot to say exactly what they want them to say, and coerced it when it deviated." <_<
-
@malwaretech@infosec.exchange "so called expert prompts a text gen bot to say exactly what they want them to say, and coerced it when it deviated." <_<
@bongmaster @malwaretech@infosec.exchange yeah tbf if you prompt an LLM "Isn't the answer yes?" it will do its best to make the answer into a yes
-
I can’t believe we’re doing this again. It’s just a bot that generates the text you ask it for. If you put it in charge of critical decisions, it will kill people. Not because it’s secretly evil, but because it’s a word generator. It’s like putting your toaster in charge of air traffic control.
-
It doesn't really matter is the glorified text generator alive or not.
It does matter how many nukes it launches if put into the position of HAL, and encountered with a situation where the matrix feels best to stop humans from shutting it down.
-
J jwcph@helvede.net shared this topic