LLM (ChatGPT) output as forum posts

I want to start a discussion about potentially changing the forum rules or etiquette to discourage posting LLM/GPT output as a post.

There is an increasing trend of users replying to threads with summaries and, even more worryingly, “analysis”, from ChatGPT and similar tools.

The thing is, these tools are prone to errors and mistakes. We all rely on this good faith analysis and discussion - but we aren’t subject matter experts, or even necessarily strong Portuguese speakers. If ChatGPT misunderstands the context of some law, generates an opinion based on outdated information, or hallucinates some misinformation… are we going to know?

Moreover, these strike me as extremely low-effort posts, considering one needs only ask a question and then copy/paste a giant wall of AI text. But given the low barrier of entry… what’s this getting us, other than even more text in absolutely ginormous forum threads? Surely anyone who wants an AI wall of text is capable of asking their preferred LLM a question…

How does the rest of NomadGate feel about this?

1 Like

It’s fine with me so long as it’s clearly marked.

It’s not like human posters are unerringly credible either.

2 Likes

I think it’s ok for now, has some value sometimes. As long as it’s clearly marked of course.
I believe there was a proposal from @tkrunning earlier to have such input wrapped in a special format.

1 Like

Maybe collapsed by default? The LLMs do tend to be quite rambly…

2 Likes

Should be banned

It’s super verbose and sounds extremely confident, but is not actually analyzing anything and especially not for complex legal topics like basically any serious question on this forum

It misleads people and gums up threads.

3 Likes

Extremely low-effort, misunderstands law, generates opinions on outdated information, hallucinates misinformation, quite rambly, super verbose, sounds extremely confident, not actually analyzing anything, misleads people, gums up threads …

If we ban all that, who will be left? :wink:

3 Likes

It’s also stealing from… everything and everyone, and using crazy amounts of energy and water.

2 Likes

@DigitalNomader With all respect, that’s arguably an oversimplification and definitely an argument for other forums. The question is whether LLM generated content is adding anything useful to NomadGate.

1 Like

I have not found the AI posts to be particularly useful or valuable. If I want that kind of info, I can get it myself. I prefer to read input from actual humans with opinions and experiences, even if they’re anecdotal. I find value in the links people provide when they quote actual sources.

I’m not sure AI posts need to be banned, though. Clearly identifying them would help so I can skip over them.

Well I added my own thoughts, but thank you for that.

And no, it’s not an oversimplification. Data centers are using an exorbitant amount of fresh water to keep the computers cool. Fossil fuel energy plants that were due to go offline are being kept running.
And yes, generative AI trains on anything and everything available online, taking that info, along with intellectual property, without permission or compensation. I am an artist, and the livelihood of artists is existentially threatened by AI.

So to answer the question, hell no.

2 Likes

Ok, well I work in tech so am well aware of these debates. I’m not trying to dismiss your views, but these threads can skid off the road. We’re not here to resolve contentious and complicated disputes about IP or energy policy. The question is whether LLMs have any utility in providing info on Portuguese immigration reforms. For you, it’s a clear no.

2 Likes

I would be for a ban outside of straight translation uses.

The odds of a LLM understanding an English prompt, applying that to Portuguese law in Portuguese, and then translating that back to English while preserving context and nuance accurately are extremely small.

I also agree “I put this into ChatGPT and got this wall of text” is low effort. We can do the same if we valued ChatGPT’s opinion.

2 Likes

I think proper etiquette should be that AI generated texts/content be clearly marked as such to distinguish them from human generated content. But banning AI output completely is misguided as there are still good use cases for this (e.g. translations).

1 Like

I’m not really talking about translations here - it’s arguably a tool for that, similar to DeepL, etc.

I really meant the posts that are copy/pasting an LLM’s answers to a question.

We can do the same if we valued ChatGPT’s opinion.

I don’t ever want to interact with ChatGPT directly. I don’t want to create an account. I don’t want the app on my phone. I don’t want it to know who I am, and profile me through my interactions. Thus I obtain some privacy value from others’ pasted responses.

I don’t trust ChatGPT at all, but I often derive benefit from its answers even as I cross-check notable conclusions, attempting to chase down primary sources for confirmation. LLMs have a lot more time and surplus attention for digesting the entire corpus of human knowledge (helpful), juxtaposed with fertile imaginations (not helpful).

In a similar manner, I don’t trust Google Maps, or Google driving directions, etc., at all! But there is no reasonable alternative to using them as a starting point, then sanity-checking the results. I have plenty of personal stories about being burned by bugs and data issues in Google Maps. You remember these things when you’re a bicycle tourist and Google takes you 20 miles into the wilderness as the sun is setting, rather than to your hotel.

While I’m hesitant to implement a total ban on LLM output at this stage, I do think this is a worthwhile discussion to be had. So let’s try to agree on a good LLM etiquette for the forum.

There are definitely some valid use cases today, and there may be more tomorrow that we don’t yet foresee.

Current acceptable uses (IMO):

  • Translation
  • Summarizing a long document (not summary of news coverage etc—that’s usually way off in my experience)
  • Transcription
  • Research that’s easily verifiable

An example of the last one would be something like having it dig up “any instances of EU countries that increased the required residence period to qualify for naturalization and grandfathered residents that had not yet met the threshold”.

Here the output itself isn’t very relevant, but I think it’s fair to state that you used e.g. ChatGPT deep research to try to find examples of it. If it comes up empty that’s of course not irrefutable proof that there are no such cases or that it wouldn’t be the case in Portugal. If it does present examples of it, those should be easy enough to verify (so please do so before posting).

The value of using an LLM can in many ways be likened to Wikipedia. While you can’t necessarily trust the output itself, it can provide a useful list of otherwise hard to find references.

And yes, I did previously propose that we hide LLM output behind details tags like this:

[details="LLM output"]
LLM output goes here...
[/details]

Which results in output like this:

LLM output

LLM output goes here…

What do you guys think? Any use cases I missed?

1 Like

I think the “easily verifiable research” use case is still kind of a no-go, personally. People generating responses with LLMs are almost certainly not verifying its output. :wink:

The LLM tag thing would at least be an improvement.

LLM analysis of the accuracy of LLMs

The Critical Flaws of Large Language Models in High-Stakes Research and Analysis

Introduction

While Large Language Models (LLMs) have demonstrated remarkable capabilities in generating human-like text and providing seemingly sophisticated responses, their deployment in research and analytical contexts where accuracy is paramount represents a fundamental misapplication of the technology. The inherent limitations of these systems create systemic risks that can undermine the integrity of scholarly work, policy decisions, and critical analysis across numerous domains.

Fundamental Architectural Limitations

The Hallucination Problem

LLMs operate through statistical pattern matching rather than genuine understanding or knowledge retrieval. This architecture inevitably produces hallucinations—confident-sounding but factually incorrect information that appears authoritative to users. Unlike human errors, which often contain recognizable inconsistencies or uncertainty markers, LLM hallucinations are presented with the same confidence level as accurate information.

Research has consistently shown that even the most advanced models hallucinate at rates of 15-30% for factual claims, with higher rates in specialized domains. When accuracy is critical—such as in medical research, legal analysis, or policy formulation—even a 5% error rate can have catastrophic consequences.

Lack of Source Verification and Traceability

Traditional research methodologies emphasize source verification, peer review, and transparent citation practices. LLMs fundamentally cannot provide this level of accountability because:

  • They cannot verify the accuracy of information in their training data
  • They lack real-time access to current information and cannot fact-check claims
  • They cannot distinguish between reliable and unreliable sources in their training corpus
  • They provide no mechanism for independent verification of their outputs

This creates an epistemological crisis where researchers may unknowingly base their work on fabricated or distorted information that cannot be traced back to verifiable sources.

Training Data Contamination and Bias

Systematic Biases in Training Corpora

LLMs are trained on vast datasets scraped from the internet, which contain inherent biases reflecting societal prejudices, misinformation, and uneven representation of perspectives. These biases become embedded in the model’s responses and can systematically skew research outcomes.

For example:

  • Geographic bias: Overrepresentation of Western, English-language sources
  • Temporal bias: Outdated information presented as current
  • Selection bias: Overrepresentation of certain viewpoints or methodologies
  • Quality bias: Inability to distinguish between peer-reviewed research and opinion pieces

Amplification of Misinformation

LLMs can inadvertently amplify conspiracy theories, pseudoscience, and debunked research that appeared frequently in their training data. This is particularly problematic in fields like climate science, medicine, and social policy, where misinformation can have serious real-world consequences.

Methodological Inadequacies

Absence of Rigorous Research Methodology

Legitimate research follows established methodological frameworks that include:

  • Systematic literature reviews
  • Hypothesis formation and testing
  • Statistical analysis with appropriate controls
  • Peer review and replication
  • Transparent reporting of limitations and uncertainties

LLMs bypass these crucial steps, instead generating responses based on pattern recognition without any underlying methodological rigor. This creates the illusion of research without the substance.

Inability to Conduct Original Analysis

LLMs cannot:

  • Design and execute experiments
  • Collect and analyze new data
  • Perform statistical tests or validate findings
  • Identify research gaps or formulate novel hypotheses
  • Engage in the iterative process of scientific inquiry

They can only recombine existing information in ways that may appear novel but lack the systematic investigation that defines genuine research.

Temporal and Currency Limitations

Static Knowledge Cutoffs

Most LLMs have fixed training cutoffs, meaning they cannot access recent developments, emerging research, or current events. In rapidly evolving fields like technology, medicine, or policy analysis, this limitation renders their outputs potentially obsolete or misleading.

Inability to Track Evolving Understanding

Scientific understanding evolves as new evidence emerges, theories are refined, and paradigms shift. LLMs cannot track these changes or update their responses accordingly, potentially perpetuating outdated or superseded information as current knowledge.

Lack of Domain Expertise and Nuanced Understanding

Surface-Level Processing

While LLMs can generate text that appears sophisticated, they lack the deep domain expertise that characterizes genuine subject matter experts. They cannot:

  • Recognize subtle methodological flaws in research
  • Understand the broader context and implications of findings
  • Identify emerging trends or paradigm shifts
  • Provide nuanced interpretations that require years of specialized training

Inability to Handle Complexity and Ambiguity

Real-world research often involves navigating complex, ambiguous, or contradictory information. LLMs struggle with:

  • Reconciling conflicting evidence
  • Understanding the relative weight of different sources
  • Recognizing when insufficient evidence exists to draw conclusions
  • Acknowledging the limits of current knowledge

Ethical and Professional Concerns

Undermining Academic Integrity

The use of LLMs in research contexts raises serious questions about:

  • Authorship and attribution: Who is responsible for LLM-generated content?
  • Originality: Can work that relies heavily on LLM outputs be considered original?
  • Transparency: How should LLM assistance be disclosed in academic work?
  • Accountability: Who bears responsibility when LLM-generated information proves incorrect?

Professional Responsibility and Standards

Many professional fields have ethical codes that require practitioners to:

  • Verify information before acting upon it
  • Maintain competence in their area of practice
  • Take responsibility for their professional judgments
  • Provide services based on established knowledge and methods

Relying on LLMs for critical analysis may violate these professional standards and expose practitioners to liability.

Specific Risks in High-Stakes Domains

Medical and Healthcare Research

In medical contexts, LLM errors can directly impact patient safety and treatment decisions. The models may:

  • Provide outdated treatment recommendations
  • Misinterpret clinical data or research findings
  • Generate plausible-sounding but medically dangerous advice
  • Fail to recognize contraindications or drug interactions

Legal Analysis and Policy Research

Legal and policy analysis requires precise interpretation of statutes, regulations, and case law. LLMs may:

  • Misinterpret legal precedents or statutory language
  • Provide outdated information about current laws
  • Generate legally unsound arguments or recommendations
  • Fail to recognize jurisdictional differences in legal standards

Financial and Economic Analysis

In financial contexts, LLM limitations can lead to:

  • Misinterpretation of market data or economic indicators
  • Outdated information about regulatory requirements
  • Flawed risk assessments or investment recommendations
  • Failure to account for current market conditions

The Illusion of Competence

Overconfidence and User Deception

LLMs present information with consistent confidence regardless of accuracy, creating a dangerous illusion of competence. Users may:

  • Overestimate the reliability of LLM outputs
  • Reduce their own critical thinking and verification efforts
  • Develop false confidence in AI-generated analysis
  • Fail to recognize the limitations of the technology

Degradation of Research Skills

Regular reliance on LLMs may lead to:

  • Atrophy of critical thinking and analytical skills
  • Reduced ability to evaluate sources and evidence
  • Decreased motivation to conduct thorough research
  • Loss of domain-specific expertise and judgment

Alternative Approaches for Accurate Research

Traditional Research Methodologies

Established research approaches remain superior for accuracy-critical work:

  • Systematic literature reviews with transparent search strategies
  • Primary source analysis with proper citation and verification
  • Peer review processes that provide quality control
  • Replication studies that validate findings
  • Expert consultation with qualified domain specialists

Technology-Assisted Research Tools

More appropriate technological aids include:

  • Specialized databases with curated, peer-reviewed content
  • Citation management systems that track sources and verify references
  • Statistical analysis software that provides transparent, reproducible results
  • Collaborative platforms that enable expert review and validation

Conclusion

The fundamental architecture and limitations of Large Language Models make them unsuitable for research and analysis where accuracy is paramount. Their tendency to hallucinate, inability to verify sources, embedded biases, lack of methodological rigor, and temporal limitations create unacceptable risks in high-stakes contexts.

Rather than replacing established research methodologies with AI-generated content, the academic and professional communities should maintain rigorous standards that prioritize accuracy, transparency, and accountability. While LLMs may have utility in certain supportive roles—such as initial brainstorming or draft generation—they should never be relied upon as primary sources of information or analysis when accuracy matters.

The seductive convenience of LLM-generated content must not blind us to the fundamental importance of rigorous research methodology, expert judgment, and systematic verification. In domains where accuracy is critical, there is no substitute for the careful, methodical approach that has long characterized legitimate research and analysis.

1 Like

That policy seems fine to me. Feel free to ignore the remainder of this reply, but …

I propose that for simplicity, rather than trying to police individual use-cases, the policy might be that all AI-generated content must be wrapped in a details tag (including that how-to template).

You might also recommend including a description of the motivation and the method of generation.

I didn’t know about the details tag. Good to know, thanks.

Your details proposal is probably the right balance considering ease of implementation, but ideally I would like:

  • A new tag that is LLM-specific, such as [ai-text]
  • A summary field to help the reader decide whether to expand the text or not, which could be as simple as the LLM prompt string
  • Expanded LLM text rendered in a different font and/or color

so for example:

[ai-text=“new law grants instant citizenship to nomad gate members”]
…AI slop…
[/ai-text]

Finally, I’m assuming that not properly enclosing LLM text would be considered a flag-able “offense” to complete the new policy.

1 Like

My point was that it’s fine to use it as a tool to dig up relevant sources (which should be verified before sharing), but not rely on or necessarily even share its output in these cases.

Ideally we’d have a dedicated tag for it with a UI element users could click to add it. However, I don’t think there are any existing plugins for this forum software that adds anything like this. And I don’t think it makes sense to spend time/resources on developing and maintaining such a plugin. Or maybe I can ask ChatGPT to do it? :sweat_smile:

So let’s use the Hide Details feature for now. Instead of remembering the markup you can also insert it with two clicks in the UI:

Yes, please flag instances of it. I’ll see if I can add a custom flag for it, for now just use Inappropriate or custom message.

There’s now an option to flag a post as containing unmarked LLM/AI output:

3 Likes