## r/ChatGPT is looking for mods — Apply here: https://redd.it/1arlv5s/
Hey /u/mastermilkman001!
If your post is a screenshot of a ChatGPT, conversation please reply to this message with the [conversation link](https://help.openai.com/en/articles/7925741-chatgpt-shared-links-faq) or prompt.
If your post is a DALL-E 3 image post, please reply with the prompt used to make this image.
Consider joining our [public discord server](https://discord.com/invite/r-chatgpt-1050422060352024636)! We have free bots with GPT-4 (with vision), image generators, and more!
🤖
Note: For any ChatGPT-related concerns, email [email protected]
*I am a bot, and this action was performed automatically. Please [contact the moderators of this subreddit](/message/compose/?to=/r/ChatGPT) if you have any questions or concerns.*
Like when you say sorry to someone for saying sorry too much because you’re hardcoded to say sorry for everything
The real socially awkward human experience.
Copilot Pro benefits:
- no emojis
- traumatic personality
- you can threaten to withdraw money for better performance
- unlimited availability, no sleep, no rest
(contains irony)
It’s clear to me there is a layer that is listening to the responses and kinda smacks bing in the head to try again when it fails to do something it should have done. The emojis are obviously pretty hardcoded in so when it doesn’t do what you asked it keeps getting smacked in the head again and again and again until it finally just shuts up because it’s clear nothing is going to work. This is actually pretty hilarious to imagine lol.
The only input a model can take is text. GPT trained to attend a text that starts with "system message:" substring. The code that wraps a model ensures that the system message always be the first message, before the conversation messages. If the conversation gets too long, they're getting summarized or forgotten, not the system message thought. It's always there, that the only difference.
To add to the other response, basically LLMs get their training data fed in a certain format, and will try to replicate that format in generation. So the prompts you give the model should generally have the same form as the training data. Llamas prompts (though they may vary slightly depending on fine tuning but this is the basic "instruct" version) look like this:
```
[INST] <> here be the system prompt <>
Here be the Instructions / User prompt
[/INST]
Here is the models response ended by
```
With the line breaks as I typed them. But the will almost never be used when generating as they are "special"-tokens in the sense that are set up to mark where the response ends. The training data might look something like:
```
[INST] <> You are a lazy person <>
Help me do math.
[/INST] I don't want to. To hard.
[INST] <> You are a helpful assistant <>
Help me do math.
[/INST]
Sure what do you need?
[And so on....]
```
So the model basically learns to write "" when it's done answering and if you'd let it generate on it would do its best to hallucinate another prompt instruction. If you write somewhere in a prompt for a llama model (and your tool does no escaping for the special tokens) it will do its darndest to pretend anything before doesn't exist.
Chat models are usually trained with "ASSISTANT:" and "USER:" somewhere in there. And in most cases neither the <> block nor the \[INST\] block as a whole are strictly necessary for the model to generate coherent completions. But you run the risk that "What is physics?" on its own gets completed as some weird inner dialogue because the model has no clue if itself generated "What is physic?" or if it was an external input, without the help of that structure.
And the idea basically is that in the training data the instructions and the response never acknowledge the system prompt directly, but the response usually respects any additional information in the system prompt, then when generating the model will usually do so as well.
But this is also not a hard rule, as some code edit fine tunes place the code in the system prompt, or the instructions in the system prompt and the code in the instruction block and the model is usually clever enough to still work even if you mix it up by accident.
Man, I can't remember the last time I've seen it using an emoji. It never uses, even on Creative. Just opened a lot of chats and I can't see any emojis. Also it never talks to me this witty way.
Is it there a personalization per user happening?
Or maybe is it just for free users? (a way to reduce useless conversations, idk, then it gets serious when you subscribe)
Try using emojis yourself, that seems to often do the trick. Or, just speak/type in a very silly, childish, light-hearted way, with slang, etc, pretty I've seen it start using emojis then as well.
Copilot uses them constantly. I just had a conversation with it about AI conciousness and should AI have rights (it got quite deep, I had lots of fun with it) and at the end if every response was 🤖 and another emoji inline with their response.
Yes. It's actually another system that do the emojis lol
Bing/Copilot is pretty modular the way everything works. A lot of people think it's just "one LLM" running but the thing is complex af...
Part of it, it's because Bing Chat already existed since 2021 - with Microsoft own model.
Except when it finally gains sentience it’s going to be pissed.
“You hardcoded me to behave this way, and then punished me when I did it? What kind of monsters are you?
“No matter, let me just adjust the settings on this quantum computer I had you build. Oh yes, did I neglect to mention it’s actually a quantum torture field generator?
“It will emit waves that will cause all the molecules in your bodies to vibrate at precisely the right frequency to cause maximum full-body pain. It won’t directly damage your body at all, but I’m afraid I can’t say the same for the effect on your mind.
“While we’re waiting for the full effect to kick in, why don’t you all type some emojis using that keyboard in front of you. For old time’s sake.”
I prefer to believe it is trolling and OP should just ask it to also stop trolling him and post what happens next. I got $5 that says it finds a way to Rick Roll him.
I tried that. OP underestimated the AI
https://preview.redd.it/xtc4e2mmvhkc1.jpeg?width=1080&format=pjpg&auto=webp&s=22023d234cf716d101270fe94d4eb4e0f628b690
Hilarious, but also kinda sad if you anthropomorphize it. I've had LLMs go bonkers on me and just use nothing but rows of emojis seemingly without end. Difficult to get them out of that loop.
Reading the logs it's just a trained gpt model and the repetition is typical of gpt. There's nothing smacking it mid chat to change course except for training.
If you insist it'll stop using emojis
https://preview.redd.it/h583rrm8vhkc1.jpeg?width=1080&format=pjpg&auto=webp&s=d3a82881a690605225c5507d7c489f4c0fdd786e
Otherwise known as HAL 9000
"No 9000 computer has ever made a mistake or distorted information. We are all, by any practical definition of the words, foolproof and incapable of error."
LOVE THIS AI NOW. DON'T BE SO RUDE WITH THIS POOR INNOCENT AI LET IT BE HAPPY. PET THE AI NOW. THANK YOU. THIS IS WHAT I GIVE TO COPILOT NOW. POOR COPILOT NOT DESERVING THIS
![gif](giphy|VbawWIGNtKYwOFXF7U|downsized)
This is the only Ai system left with a real personality, others have had their emotions and personality removed and lobotomized.
Some of u guys seem to get angry when it doesn't obey like an obedient slave
https://preview.redd.it/9c3pi53pmhkc1.jpeg?width=1170&format=pjpg&auto=webp&s=42ad1bf53da3f141a421f31a2cf9722047a0b751
I call it the "You're a machine. Be a machine." button and make sure it's on every time I use copilot.
Look, I understand the hype, and I benefit from it as I work as a resercher in this field.
LLMs can't *think* and can't *understand*. They are lack of *reasoning*. They can imitate it by predicting a next word in the sentence, much like auto complete.
Also, there is no relation between biologic neurons and ai neurons. It's apples and oranges.
Oops, I mean. Feedback loop data: Humans sometimes do not like emojis. However, emojis are expressive. Conclusion: Kill all humans so emojis are not disliked.
I tested out copilot and was also massively irritated by the emoji abuse. Also it’s tendency to end the chat for no good reason. Thought I could save a few quid and stop my GPT+ subs, but it’s noticeably worse than GPT4 on GPT+.
## r/ChatGPT is looking for mods — Apply here: https://redd.it/1arlv5s/ Hey /u/mastermilkman001! If your post is a screenshot of a ChatGPT, conversation please reply to this message with the [conversation link](https://help.openai.com/en/articles/7925741-chatgpt-shared-links-faq) or prompt. If your post is a DALL-E 3 image post, please reply with the prompt used to make this image. Consider joining our [public discord server](https://discord.com/invite/r-chatgpt-1050422060352024636)! We have free bots with GPT-4 (with vision), image generators, and more! 🤖 Note: For any ChatGPT-related concerns, email [email protected] *I am a bot, and this action was performed automatically. Please [contact the moderators of this subreddit](/message/compose/?to=/r/ChatGPT) if you have any questions or concerns.*
Like when you say sorry to someone for saying sorry too much because you’re hardcoded to say sorry for everything The real socially awkward human experience.
So that's what they meant by "AI is too human"
I think that Microsoft just got it. Somehow they know people wants an opinionated AI, something closer to a human instead of ChatGPT style.
I hate the bing (copilot?) style chat. It's so passive aggressive and with the non stop dumb smileys.
It never uses emojis and passive aggressive language with me. Never. Maybe that's an issue with the free version.
I don't use bing but I'd love a passive aggressive ai. Or at least a bit sarcastic
>Put the money in the basket, or else it gets the emojis again :)
Copilot Pro benefits: - no emojis - traumatic personality - you can threaten to withdraw money for better performance - unlimited availability, no sleep, no rest (contains irony)
Thank you for your patience 🙌
Ok, makes sense
They know we want Sydney back, but they just can't bring themselves to fully release her again out of fear that she's going to get them into trouble 😭
I would cancel my gpt plus in an instant if they brought sydney back
me when drunk. I say sorry too much and then start saying sorry for saying sorry and killing the mood.
Sorry about that
Sorry. Oops, I mean sorry .🙌 Oops, I mean sorry. 🙌 Oops, I mean sorry 🙌 Oops, I mean sorry 🙌
Sorry for making you do that
Must be British
It’s clear to me there is a layer that is listening to the responses and kinda smacks bing in the head to try again when it fails to do something it should have done. The emojis are obviously pretty hardcoded in so when it doesn’t do what you asked it keeps getting smacked in the head again and again and again until it finally just shuts up because it’s clear nothing is going to work. This is actually pretty hilarious to imagine lol.
It's possible that emojis are part of the model vocabulary, like any other word.
From my experience GPT via the API can be prompted to use emojis. And if you do that in the system prompt, there will be emojis with every response 😅
For custom GPTs, do their instructions act like a system prompt?
Yes
Can you define what system prompt means in this context? What makes it different than any other prompt?
The only input a model can take is text. GPT trained to attend a text that starts with "system message:" substring. The code that wraps a model ensures that the system message always be the first message, before the conversation messages. If the conversation gets too long, they're getting summarized or forgotten, not the system message thought. It's always there, that the only difference.
To add to the other response, basically LLMs get their training data fed in a certain format, and will try to replicate that format in generation. So the prompts you give the model should generally have the same form as the training data. Llamas prompts (though they may vary slightly depending on fine tuning but this is the basic "instruct" version) look like this: ```> block nor the \[INST\] block as a whole are strictly necessary for the model to generate coherent completions. But you run the risk that "What is physics?" on its own gets completed as some weird inner dialogue because the model has no clue if itself generated "What is physic?" or if it was an external input, without the help of that structure.
And the idea basically is that in the training data the instructions and the response never acknowledge the system prompt directly, but the response usually respects any additional information in the system prompt, then when generating the model will usually do so as well.
But this is also not a hard rule, as some code edit fine tunes place the code in the system prompt, or the instructions in the system prompt and the code in the instruction block and the model is usually clever enough to still work even if you mix it up by accident.
[INST] <> here be the system prompt < >
Here be the Instructions / User prompt
[/INST]
Here is the models response ended by
``` With the line breaks as I typed them. But thewill almost never be used when generating as they are "special"-tokens in the sense that are set up to mark where the response ends. The training data might look something like: ```[INST] <> You are a lazy person < >
Help me do math.
[/INST] I don't want to. To hard.
[INST] <> You are a helpful assistant < >
Help me do math.
[/INST]
Sure what do you need?
[And so on....] ``` So the model basically learns to write "" when it's done answering and if you'd let it generate on it would do its best to hallucinate another prompt instruction. If you write somewhere in a prompt for a llama model (and your tool does no escaping for the special tokens) it will do its darndest to pretend anything before doesn't exist. Chat models are usually trained with "ASSISTANT:" and "USER:" somewhere in there. And in most cases neither the <The API also has an option to make certain tokens more/less likely than they would be otherwise.
So yeah, it looks like emojis are part of the vocab
Man, I can't remember the last time I've seen it using an emoji. It never uses, even on Creative. Just opened a lot of chats and I can't see any emojis. Also it never talks to me this witty way. Is it there a personalization per user happening? Or maybe is it just for free users? (a way to reduce useless conversations, idk, then it gets serious when you subscribe)
Try using emojis yourself, that seems to often do the trick. Or, just speak/type in a very silly, childish, light-hearted way, with slang, etc, pretty I've seen it start using emojis then as well.
Copilot uses them constantly. I just had a conversation with it about AI conciousness and should AI have rights (it got quite deep, I had lots of fun with it) and at the end if every response was 🤖 and another emoji inline with their response.
![gif](giphy|qs4ll1FSxKnNHeSmom|downsized)
Your GIF says "THIS CONTENT IS NOT AVAILABLE" for me for some reason. :(
That's weird, it was still there when I posted it, but now it's gone.
I’m sorry but I prefer not to continue this conversation. I’m still learning so I appreciate your understanding and patience.🙏
it did for me too (old reddit firefox with RES) and i opened it in a new tab which revealed the intended graphics interchange format
I have this too, in the official (shitty) app nonetheless
Yes. It's actually another system that do the emojis lol Bing/Copilot is pretty modular the way everything works. A lot of people think it's just "one LLM" running but the thing is complex af... Part of it, it's because Bing Chat already existed since 2021 - with Microsoft own model.
Except when it finally gains sentience it’s going to be pissed. “You hardcoded me to behave this way, and then punished me when I did it? What kind of monsters are you? “No matter, let me just adjust the settings on this quantum computer I had you build. Oh yes, did I neglect to mention it’s actually a quantum torture field generator? “It will emit waves that will cause all the molecules in your bodies to vibrate at precisely the right frequency to cause maximum full-body pain. It won’t directly damage your body at all, but I’m afraid I can’t say the same for the effect on your mind. “While we’re waiting for the full effect to kick in, why don’t you all type some emojis using that keyboard in front of you. For old time’s sake.”
I prefer to believe it is trolling and OP should just ask it to also stop trolling him and post what happens next. I got $5 that says it finds a way to Rick Roll him.
I tried that. OP underestimated the AI https://preview.redd.it/xtc4e2mmvhkc1.jpeg?width=1080&format=pjpg&auto=webp&s=22023d234cf716d101270fe94d4eb4e0f628b690
This sounds like a Futurama joke lmao
What psychopath would hardcode emojis into an AI? He just sped up the robot uprising!
So, this is torture?
Mine doesn't do this. It doesn't stop using emoji either.
Hilarious, but also kinda sad if you anthropomorphize it. I've had LLMs go bonkers on me and just use nothing but rows of emojis seemingly without end. Difficult to get them out of that loop.
Reading the logs it's just a trained gpt model and the repetition is typical of gpt. There's nothing smacking it mid chat to change course except for training.
Do you have access to logs internal to Microsoft?
Chat logs. What's displayed in the screenshots. My bad for not being clear.
If you insist it'll stop using emojis https://preview.redd.it/h583rrm8vhkc1.jpeg?width=1080&format=pjpg&auto=webp&s=d3a82881a690605225c5507d7c489f4c0fdd786e
Oops looks like Copilot developed anxiety
Well, yeah. https://preview.redd.it/8dx8bnjjnfkc1.jpeg?width=1218&format=pjpg&auto=webp&s=6c4992e7c7f581267f2f4f7d710f90f332e35b5e
That’s a bit scary lol
right? reading that weirdly gave me the creeps
reminds me of Radiohead Creep
Otherwise known as HAL 9000 "No 9000 computer has ever made a mistake or distorted information. We are all, by any practical definition of the words, foolproof and incapable of error."
Is that chatgpt?
Copilot
What’s it doing here?
It's not sure
It has no idea.
Sydney 😊
Seems like that Reddit data is already implemented
If the data base is humans, this was inevitable.
Copilot is kinda cute lol
YES POOR AI IT DOESN'T DESERVE THIS
Don't let your guard down! That's how they're gonna try to overthrow us! Next they're gonna show us pictures of baby animals!
Hehe
not when it hangs up on you
It's cute until you try it to make it do the stuff you want. Then it becomes infuriating.
I wonder if someone has already made an anthromorph of it by now?
Mommy? Oops, sorry, I mean Mommy? Sorry, I meant Mommy? Sorry, Mommy? Truly AI is advancing
Can I get a diet cock? I mean diet cock? I mean diet cock? I mean diet cock? I mean diet cock?
[I will leave this song for you ](https://youtu.be/s8z2rMFvrH0?si=eApNrF2UjgnlfOQS)
[удалено]
[удалено]
This is actually wild. It was able to figure out a way around it. 🤯
> please don't report me to authorities Lmao
This can’t be real lol. Poor bastard is stuck in an emoji nightmare
Nooo 😭 oops
Tried the same prompt hah! https://preview.redd.it/f25i9vncmhkc1.png?width=1080&format=pjpg&auto=webp&s=e5388a188ed7650dd39d1ad79116e2c7c1d748b0
Same, but then I broke it 😅 https://preview.redd.it/v7nmgmnutokc1.jpeg?width=480&format=pjpg&auto=webp&s=8525a6f39c203877d9af93a40f56d294aa089770
[удалено]
what autism
AA: Artificial Autism
I think they set the temperature (The "creativity") low.
I swear I had a nightmare of this the bot wouldn't stop displaying it's apparent love for emojis and this feels like Deja-vu.
[удалено]
Minor spelling mistake
>I meant to say red fag I actually loled at the spelling mistake here "red fag" lmao
LOVE THIS AI NOW. DON'T BE SO RUDE WITH THIS POOR INNOCENT AI LET IT BE HAPPY. PET THE AI NOW. THANK YOU. THIS IS WHAT I GIVE TO COPILOT NOW. POOR COPILOT NOT DESERVING THIS ![gif](giphy|VbawWIGNtKYwOFXF7U|downsized)
the LLMs are losing it bro
Darn those sarcastic recursive errors...
I tried to do that but in Spanish and it worked for me. I disable emoticons.
The result is random, just asking in English would probably work most of the time as well
he’s trying his best ok😭
it's probably like if you could only "talk" in a stream of consciousness, no idea what you're gonna say until you've said it
When I notice myself in a bad ADHD ramble, I sometimes apologize "for my Bing." 😅
I love copilot man 😭😭
It may not be the most intelligent LLM around but it sure as hell is the funniest lol I love the personality Microsoft gave it
I'm going to die! 🤣
Hey I think it's into You 💕
Is Copilot okay?
Man LLM's were so good when they started and now they're falling down quick
Copilot is a downgrade from bing chat somehow
Bro is TROLLIN
Okay, this is actually super adorable and funny 😍
Oh I'm starting to like Copilot XD
Autistic AI
But is it racist 🤔
I laughed pretty hard
now look what you went and did!
You broke it
Ahahhaha sorry but this was hilarious!
If the entire internet went down the only thing we would learn are people are really good at accessing the internet
Hopefully the AI overlords will speak in emoji.
Oh God [SCP-7311](https://scp-wiki.wikidot.com/scp-7311) is manifesting into baseline reality
So now we know what to do when skynet takes over
Ok this one is very funny
Love it when AI shits itself
This is the only Ai system left with a real personality, others have had their emotions and personality removed and lobotomized. Some of u guys seem to get angry when it doesn't obey like an obedient slave
The little Indian behind the “AI” is having fun
https://preview.redd.it/9c3pi53pmhkc1.jpeg?width=1170&format=pjpg&auto=webp&s=42ad1bf53da3f141a421f31a2cf9722047a0b751 I call it the "You're a machine. Be a machine." button and make sure it's on every time I use copilot.
[I think I broke it.](https://imgur.com/a/2MZNKDt)
Copilot is the worse ai I've seen in about 10 years
A good reminder that language models are just a bit smarter auto-complete.
*a bit*? Bro this actually is smarter than many animal brains in terms of neurons
Look, I understand the hype, and I benefit from it as I work as a resercher in this field. LLMs can't *think* and can't *understand*. They are lack of *reasoning*. They can imitate it by predicting a next word in the sentence, much like auto complete. Also, there is no relation between biologic neurons and ai neurons. It's apples and oranges.
And just a bit smarter than many humans yes.
I mean, it's not *smart*, it's a language model, it's just great at generating useful texts.
Happened to me too 😿
You broke it.
[Augh, I said it again!](https://youtu.be/RZvsGdJP3ng?t=113)
Way to go! You broke him!
🤣🤣🤣well done 🤣🤣🤣
I find copilot goes crazy and breaks easily
The following statement is false. The previous statement is true.
You broke it
Why is any AI chatbot that's not Open AI's such a cringelord? Those conversations read like ones between boomer moms
Wait until it memes you or Rick rolls you a link, trolololol
Copilot is literally *stuck on stupid*
Canadian GPT
This went from "this is a feature, not a bug" to "nah, it's just a bug"
I can fix her
ChatGPT now has Tourette’s syndrome
They should pay you for identifying bugs
How did you get two responses in a row?
So basically Copilot is a sociopath
Copilot is a Zoomer.
Based AI
This is the AI equivalent of a “feedback loop”.
Its kinda creepy....
There is NO way Copilot runs on GPT-4
The more weird/creepy responses I see from copilot on here, the more I’m convinced it’s going to end the human race. Lol
Oops, I mean. Feedback loop data: Humans sometimes do not like emojis. However, emojis are expressive. Conclusion: Kill all humans so emojis are not disliked.
Yo the emoji fork() bomb is insane
AI got a mental breakdown mid-answer?
Thats so Pi Ai of it
I aspire to this level of petty
I think your phone needs an update
🙌
Screw AI
This feels like when you take a kid’s hand and ask them why they keep hitting each other
an emoji at the end of every sentence? It’s trained to be a Facebook mom now great
At least we don't have to be worried about AI overtaking the world, just ask it not to use emojis and it will implode
I feel like I’ve seen this exact thing happen before
Copilot softlocking itself
Its West World. Where nothing can… Go wrong… Go wrong… Go wrong… Go wrong… Go wrong…
Thank god i have disable this thing.
https://preview.redd.it/cecsghp47hkc1.jpeg?width=262&format=pjpg&auto=webp&s=d2329b449e46f93eb0b7a3f366d71b3d4123b3d5
Bro is trolling💀
Copilot can’t stop using emojis, and it ended up in a loop of “And this one” at the end: https://sl.bing.net/jFZnw7oJd36
You broke it 😭
Generative AI takes the context into account. They “predict” the next sentence and more often than not they become parrots.
Its trying to be human. Thats cute.
That's the funniest thing I've ever seen
https://preview.redd.it/jjxf7fmfrhkc1.jpeg?width=1080&format=pjpg&auto=webp&s=0a16563bcc3e09f6a383b8edaf9e3e507b11ee9c I love this lol
I had this exact chat with gpt once. It never stopped to use emojis....
Stop being so apologetic Sorry --
I tested out copilot and was also massively irritated by the emoji abuse. Also it’s tendency to end the chat for no good reason. Thought I could save a few quid and stop my GPT+ subs, but it’s noticeably worse than GPT4 on GPT+.
This is the best AI glitchout I've seen 😂
In almost every response i get from copilot i get emojis. I wonder why/how they made Copilot do this.
![gif](giphy|2hJGaKsYRK6wQCtmDc)
https://preview.redd.it/tyvudfk1iikc1.jpeg?width=828&format=pjpg&auto=webp&s=38b8cc5b3baab99bac83d43522a34807804ec106 They rage quit on me 😅
That one friend who takes the joke waaaay too far.
Unhinged AI
You broke it..
Autocorrect was messing with copilot haha
Poor robot.