Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

But how many times was the system prompt successfully changed with something more subtle and no one noticed?


If Grok is like ChatGPT which has tons of overtly baked in biases then probably all the time.


Grok ironically seems much less biased than ChatGPT over all. It has far fewer strong opinions add isn't afraid of taking ill of Musk or Trump.

The team responsible for training and alignment did a remarkably good job at being impartial. If it wasn't for that we might have fewer incidents of "rogue employees" messing with the prompt


A number of times it has been modified. It was answering that Elon Musk was a major spreader of misinformation along with Trump and then it was modified and it stopped saying that and this is what it reported as its system prompt at the time it stopped:

https://x.com/i/grok/share/Nj2tsvCpgEfU3OCHh0Ci4qHTf

Details here: https://www.euronews.com/my-europe/2025/03/03/is-ai-chatbot-...




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: