Community policy on AI-generated answers (e.g. ChatGPT)

Add FAQ category that is locked for moderators + trusted only to write into?

Would the topics in that category then be always viewable and linkable?

We’ll probably see an emerging use of Google Bard on this forum over the coming months. I asked Bard an autobiographical question this morning, as follows:

Me: Is Bard a large language model?

Bard: Yes, Bard is a large language model (LLM) developed by Google AI. It is a conversational AI that can understand and respond to natural language queries. Bard is trained on a massive dataset of text and code, and it can generate text, translate languages, write different kinds of creative content, and answer your questions in an informative way.

LLMs are a type of artificial intelligence (AI) that can generate text, translate languages, and write different kinds of creative content. They are trained on massive datasets of text and code, and they can learn to mimic the style and structure of human writing. LLMs are still under development, but they have the potential to revolutionize the way we interact with technology.

This morning, I thought to throw a somewhat naive question about Python to ChatGPT and Bard in order to consider whether a general policy toward all large language models could cover the entire range of sophistication. ChatGPT is allowing me to retrieve both the prompt and response from this morning as text, but I can only retrieve the prompt from Bard. So, below is a text copy of the episode with ChatGPT and a screen capture of such with Bard.

First, ChatGPT

Me: If I use Python to add two binary numbers, will it give me a binary number as a result?

ChatGPT: Yes, if you use Python to add two binary numbers, the result will also be a binary number. This is because Python treats binary numbers as strings of 0s and 1s, just like any other string. When you add two binary numbers in Python, the interpreter will concatenate the two strings and perform the binary addition operation, resulting in a new binary string that represents the sum of the two original binary numbers.

For example, if you add the binary numbers “1101” and “1010” in Python using the “+” operator, the result will be the binary number “101111”. Here’s an example Python code snippet that demonstrates binary addition:

makefile

binary1 = "1101"
binary2 = "1010"

# Convert binary strings to integers and add them
decimal_sum = int(binary1, 2) + int(binary2, 2)

# Convert decimal sum back to binary string
binary_sum = bin(decimal_sum)[2:]

print("Binary sum:", binary_sum)

This code will output: Binary sum: 101111 .

Now, Bard with the same prompt …

It was somewhat funny to see that ChatGPT’s response began with “Yes,” and Bard’s with an opposite “No,”. That is not so surprising though, considering that the prompt was crafted to be a bit foggy conceptually. ChatGPT’s answer does muddle things when it brings up concatenation.

Though the above differ, a single uniform policy toward LLMs, in general, could cover both, for example this suggestion posed earlier:

Let’s recall that some of the participants in this discussion have voiced opposition to allowing the use of LLMs at all here. We’d best remain respectful of their feelings, perhaps by not being too quick to post a permissive policy.

My original view regarding the issue was:

This situation is quite nuanced and fluid, and now I’m fundamentally unsure how to proceed. By and large, the other participants in this discussion are more central to this community than myself. So, hereafter I might just ask a question now and then, and ultimately cast my view in favor of whatever consensus the community may reach.

A small aside, you did give ChatGPT and Bard two different prompts.

Your ChatGPT prompt:

Me: If I use Python to add two binary numbers, will it give me a binary number as a result?

And your Bard Prompt:

image

Notice the “binary numbers” vs. “primary numbers”.

According to Wikipedia “Prime Powers” are sometimes called “Primary Numbers”: Prime power - Wikipedia, I have no idea if that’s what Bard interpreted your question as but a better response probably would have been to ask for clarity on your question.

1 Like

Oops …


This time, we got a "Yes, ".

Hmmm … worth following my own advice.

1 Like

I’ve got a suggestion: instead of trying to detect and ban, deincentivize replying with ChatGPT by directly providing a function to call a ChatGPT answer on any post.

This may not be a practical solution right now because ChatGPT is proprietary and costly, but when there is a lot of work underway to try to match the performance with an opensource framework, and it’s only a matter of time before distilled, reduced models that can run very fast or even in the user’s browser JS interpreter to be available.

Then, it will become a very practical possibility to just offer a “AI answer this” button for any post.

Given that, as already discussed, LLMs lack any sense of factuality
for their responses, it seems like it could be very misleading to
users. People might click the “get an AI answer” button and follow
its advice, not realizing that it’s not actually an answer but
rather predictive text made to sound like a believable answer
without any significant probability of being correct.

3 Likes

Even if it might, or might not, suppress the frequency of direct incorporation of ChatGPT responses into posts on this forum, wouldn’t this proposed feature effectively promote and even appear to advocate the naive practice of uncritical solicitation and utilization of advice from large language models (LLMs)?

1 Like

Unfortunately, misplaced uncritical beliefs of users in LLMs will happen regardless of what forums officially choose, due to the automatism and anthropomorphism biases. I have worked in the past on military drones, and these biases have been well documented there already: even trained military crews, who are highly trained, have been reported to risk their lives to carry back damaged mine clearing robots in a dangerous field.

I may have a bleak outlook on the topic, but as a AI researcher, I see no way non-AI researchers can have a good enough grasp of such a highly intricate and polymorphic topic as to not unconsciously get their perception of reality biased by AI models that are now ubiquitously available (and even before ChatGPT, there were algorithmic bubbles everywhere online due to recommendation algorithms). There are already tragic examples, such as a belgian health researcher ending their life (hence someone with a fair bit of education in a developed country) because of a GPT-J based chatbot.

What I am suggesting is to commoditize these tools so that humans can learn to tame them, for example with a clear explanation of what they do, and with a clear UI distinction between AI-generated answers and real human answers. I am not suggesting as you seem to imply that AI-generated answers would just be posted on the forum, but rather suggested in a very distinct UI field, eg, like the GitHub system of recommending duplicate answers when creating an issue, it could also suggest a succinct AI-generated answer. In fact this would be a very similar UI experence to summarization algos.

Yes, the lack of a concept of correctedness is a major issue, there is no formal guarantee, but other algorithms such as summarization algorithms are widely used (even here) and they also lack such a concept.

And while I certainly agree that the lack of formal truthiness guarantees are a major issue, and especially because the bot lacks any way to search for up-to-date information, and I’m sure there are additional heuristics that can be devised to improve practical accuracy, let’s not forget that there is no way to guarantee that any statement is true. Bots cannot solve a fundamentally impossible problem that humans are not exempt of either. And this is not the first criticism that was heard about the potential for a new technology to propagate misinformation: I remember the 1990-2000s when Internet was considered a cesspool of only amateurish-fake-conspirationist information (which partially true), and the invention of the printer was likewise criticized as it could be used to print pamphlets in huge volumes.

The TL;DR of my suggestion is that, while I certainly agree with a blanket ban in the current situation given we lack tools and hindsight, progress cannot and should not be stopped, LLMs are here to stay, future technos will be even better at mimicking human discourse, so I think that the only sustainable long-term solution is to educate humans about how to use these tools and their limitations, just like what happened historically with other new communication technologies. Devising standards about how to present such AI-generated content will certainly help in this endeavor.

1 Like

On the contrary, the following implies that your proposed feature might or might not succeed in decreasing the frequency of AI-generated answers on the forum:

Please clarify whether or not the provision of the function alluded to in the following would be offered in order to make it convenient for a user to seek an AI-generated answer in lieu of their soliciting a reply directly from a human on this forum:

Reviving this a bit, it seems as predicted LLM (specifically ChatGPT) based answers are starting to proliferate. Reviewing recent #users questions and answers over the past few days, I noticed we have one new user, @Fusion9334 , who has joined 3 days ago and has answered a number of questions in a pattern which appeared to me to be very likely to be at the very least heavily assisted by ChatGPT or a similar LLM. This conjecture was apparently confirmed by checking their profile, where the one topic they have posted is a question specifically about using ChatGPT’s Python API. In fact, as it was their first post, it seems that was in fact what brought them to this forum.

Now, to be clear, this user hasn’t done anything against the rules—however, I think their history might be a very useful set of datapoints to learn and discuss more practically about how LLMs might be used (for good or ill) in answers, and how we might address any negative impacts. Furthermore, their input here would be appreciated as well.

What follows is my personal initial impressions. It seems they’ve taken at least some care to not just jump the LLM output, and tailor it somewhat to the situation faced by each poster. Additionally, it seemed like the LLM’s wide range of background allowed it to potentially answer in much more detail (if not entirely correctly) to specialized subject matter questions about specific tools and services that the typical users here helping others were unlikely to know about.

On the other hand, on many more basic Python questions (more in scope for this forum), I did notice many instances where the other posters were trying to engage the user, typically a beginning learner and often working on an assignment, in a pedagogically motivated discourse to actually help them learn. However, before I even considered a LLM might be involved, I noticed the replies by this user were running somewhat contrary to that, directly giving the a bunch of code (that may or may not address the real problem) with relatively minimal explanation, that al least implicitly encouraged them to just copy/paste rather than actually learn something, like the other people were trying to each them. In fact, I was considering mentioning it privately to the user to consider in their approach to answering future questions, before I suspected an LLM.

After examining this real-world history, I’d like to hear others’ further thoughts, discussion points and proposals. Thanks!

3 Likes

Agreed. I wasn’t thinking LLM when I saw this thread but it definitely isn’t the sort of post that we want to be encouraging. We do NOT want people taught to copy and paste code, especially when that code has come from a language model with no concept of correctness.

2 Likes

As another update, we had to perma-ban the original user who’s LLM-generated response initially prompted this thread, as their further posts were just copy-pasting other peoples’ old (and often extremely outdated) questions from Stack Overflow with an added link to a spammy site they were evidently promoting, with moderate (possibly LLM-induced) paraphrasing to attempt to obfuscate it and no attribution. Of note, this demonstrates another form of re-using external content without attribution of which the above-proposed policy would address.

Additionally, we’ve had at least one further substantial instance of a new user employing an AI, this time in the process of writing a question asking for code review, as well as subsequent followups. Specifically, they used codepal.ai to edit the code and asked ChatGPT questions about the code/responses as well as using it to guide code generation in later replies. To note, they’d previously posted another code review thread with their own code, which developed into a long and fruitful back and forth discussion/Q&A, but this one swiftly digressed into debate about the use of LLMs and similar tools in generating, editing and reviewing code:

1 Like

My fear is that we may be seeing on this forum only the tip of a large iceberg of students who are beginning to rely on AI to generate code. If this is allowed to continue, we might wind up with a huge population of programmers who are unable to program.

In response, should we just send away learners who arrive at our doorstep with code generated by AI? Perhaps that would keep our own house clean, while allowing the problem to worsen out there in the larger world.

This may be a time in history when we should try to form an alliance with OpenAI and other such outfits in a quest to urge students to learn to program by doing the hard work of planning, writing, testing, and refining code. In the long run, this would be best for all concerned.

Surely this is a self correcting situations. Programmers by definition can program.
If the AI cannot then it is not going to work for people looking for a short cut.
No more then cut-n-paste from stackoverflow makes you a programmer.

1 Like

This quote has been floating around the Internet for a while attributed to J.A.N Lee of Virginia Polytechnic Institute:

One of von Neumann’s students at Princeton recalled that graduate students were being used to hand assemble programs into binary for their early machine. This student took time out to build an assembler, but when von Neumann found out about it he was very angry, saying that it was a waste of a valuable scientific computing instrument to use it to do clerical work.

Assuming this was John von Neumann’s position he was in a way correct, how many of can hand write binary programs to directly feed the computers anymore?

Companies and programs that can spend their resources more efficiently to solve their problems ultimately win out. LLM assisted programming might be a dead end, or in 60 years not using whatever LLM evolves in to could look like hand writing binary.

Making statements right now on what we have to force students to do really ignores the arc of history of how coding has evolved.

Personally I believe this current state of LLM theory means they won’t be great at direct code generation, but may significantly help with boilerplate code, unit tests, PR reviews, and other areas that already involve existing code that needs to be extended in some way. But my views will be washed away from this tide of history no doubt.

Yet our advice to learners today must necessarily be guided by the present state of the art and science of computing. An assembler that functioned properly in von Neumann’s lab compares favorably over a present-day LLM that is not so reliable at writing code. The question at hand today is how to respond to those who approach us with solutions that are authored by AI. Our response today may differ from what it will be in the future.

1 Like

My conclusion is to avoid using AI if I am planning to ask for Code Review :slightly_smiling_face:

To summarize my thoughts, the main reason I use AI a lot is probably because I’m afraid of giving someone a bad impression, which has happened many times. Unlike AI, which I needn’t worry about.

And I’m not afraid of AI, as long as I keep an open mind, I don’t see it as threatening.

1 Like

Seems to be a wise conclusion :slight_smile: Code review of AI-generated (or at least substantially-assisted) code would tend to give feedback about where the AI might have went wrong, rather than suggestions and pointers on how you can improve your code, which seems much more useful if you’re looking to learn and improve from it (as it seems you are).

That’s actually a quite interesting point, one I at least hadn’t previously considered. It can be genuinely quite intimidating and scary at first as a beginner to ask questions and present your code for review and critique by other human experts. And I’m certainly someone myself who can have a hard time sharing with others something that I haven’t yet mastered, ever since I was little up till the present day.

We all were newbies at some point, and we all made (and continue to make) mistakes in our code sometimes. Other folks here won’t (or at least, shouldn’t) look down on you or think badly of you just for making mistakes, the same ones we all did at some point. And while I know it can be easy for me to say this and hard for you do, don’t put yourself down over making one, either—instead, maybe treat each one as a learning opportunity to further improve your skills :slight_smile:

What earns my deep respect and admiration (and likely most others here as well) is someone who, regardless of current skill level or mistakes, always tries to be open to listening, learning, improving and asking questions. And no one here should judge you negatively as a person due to making a beginner programmer mistake. We’re here to help you along in your journey as we likewise travel on our own :blush:

And if there’s something we could do, individually or as a community, to help you feel more welcome (or addressing someone or something that’s making you feel otherwise), we’d love to hear from you—please do reach out either publicly, or privately to the mod team or me directly. Thanks!

6 Likes