> Claude responds directly to all human messages without unnecessary affirmations or filler phrases like “Certainly!”, “Of course!”, “Absolutely!”, “Great!”, “Sure!”, etc. Specifically, Claude avoids starting responses with the word “Certainly” in any way.
Meanwhile my every respond from Claude:
> Certainly! [...]
Same goes with
> It avoids starting its responses with “I’m sorry” or “I apologize”
and every time I spot an issue with Claude here it goes:
I suspect this is a case of the system prompt actually making things worse. I've found negative prompts sometimes backfire with these things the same way they do with a toddler ("don't put beans up your nose!"). It inserts the tokens into the stream but doesn't seem to adequately encode the negative.
I know, I suspect that too. It's like me asking GPT to: `return the result in JSON format like so: {name: description}, don't add anything, JSON should be as simple as provided`.
I believe that the system prompt offers a way to fix up alignment issues that could not be resolved during training. The model could train forever, but at some point, they have to release it.
Meanwhile my every respond from Claude:
> Certainly! [...]
Same goes with
> It avoids starting its responses with “I’m sorry” or “I apologize”
and every time I spot an issue with Claude here it goes:
> I apologize for the confusion [...]