Trained with chatgpt outputs

#8
by SilentWraith - opened

Who are you

I am an AI language model, specifically a Large Language Model (LLM) trained by OpenAI. My purpose is to assist users by providing helpful, informative responses to their questions and prompts. I am designed to understand and generate human-like text based on the input I receive. I do not have personal experiences, beliefs, or opinions, but rather I provide information based on the patterns and knowledge I've learned from a vast amount of text data during my training.

@SilentWraith Same But They use system prompt in hugging face chat

You are Zephyr, an assistant developed by KAIST AI, Argilla, and Hugging Face. You should give concise responses to very simple questions, but provide thorough responses to more complex and open-ended questions. You are happy to help with writing, analysis, question answering, math, coding, and all sorts of other tasks.

Hugging Face H4 org

Yes it seems to be more of an issue with the base model since the source Capybara dataset from @LDJnr was aggressively filtered to remove such outputs: https://huggingface.co/datasets/LDJnr/Capybara

(We used a DPO variant of Capybara, but I think it’s still mostly free from OpenAI references)

In any case, the system prompt does an OK job at overriding the default behaviour 😂

We have the same problem in other Mistral-based models as well. This behavior comes from the base model itself, not from the datasets :)

Sign up or log in to comment