this post was submitted on 14 Jul 2023
241 points (94.1% liked)

Technology

59298 readers
4777 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
top 50 comments
sorted by: hot top controversial new old
[–] dan1101@lemmy.world 96 points 1 year ago* (last edited 1 year ago) (1 children)

As expected, they can't be trusted. And the more AI evolves, the less likely AI content will be detectable IMO.

[–] jocanib@lemmy.world 12 points 1 year ago (7 children)

It will almost always be detectable if you just read what is written. Especially for academic work. It doesn't know what a citation is, only what one looks like and where they appear. It can't summarise a paper accurately. It's easy to force laughably bad output by just asking the right sort of question.

The simplest approach for setting homework is to give them the LLM output and get them to check it for errors and omissions. LLMs can't critique their own work and students probably learn more from chasing down errors than filling a blank sheet of paper for the sake of it.

[–] weew@lemmy.ca 55 points 1 year ago (10 children)

given how much AI has advanced in the past year alone, saying it will "always" be easy to spot is extremely short sighted.

[–] Kara@kbin.social 26 points 1 year ago (1 children)

People seem to grasp onto weaknesses AI has now and say that they will have them forever, like how text AI lies, and image generation AI can't draw hands.

But these AIs are advancing unimaginably quick, 2 years ago generated text was pretty bad, becoming pretty incoherent, and 1 year ago generated images were mostly strange mush.

[–] aebrer@kbin.social 3 points 1 year ago

Spot on! Actually people still talk about hands but it's already been solved with many newer image gen models... The hands they produce look perfectly fine usually these days.

load more comments (9 replies)
[–] Zeth0s@lemmy.world 24 points 1 year ago (1 children)

This is not entirely correct, in my experience. With the current version pf gtp-4 you might be right, but the initial versions were extremely good. Clearly you have to work with it, you cannot ask for the whole work

[–] jocanib@lemmy.world 4 points 1 year ago (1 children)

That's not true! There's heaps of early-GPT articles pointing out how much bullshit it regurgitates (eg Why does ChatGPT constantly lie?). And no evidence at all that the breathless fanboys have even stopped to check.

[–] Zeth0s@lemmy.world 7 points 1 year ago* (last edited 1 year ago)

I meant initial versions of chatGTP 4. ChatGTP isn't lying, simply because lying implies a malevolent intent. Gtp-4 has no intent, it just provides an output given an input, that can be either wrong or correct. A model able to provide more correct answers is a more accurate model. Computing accuracy for a LLM is not trivial, but gpt-4 is still a good model. User has to know how to use it, what to expect and how to evaluate the result. If they are unable to do so it's completely their fault.

Why are you so pissed of a good nlp model?

[–] nulldev@lemmy.vepta.org 8 points 1 year ago (2 children)

LLMs can't critique their own work

In many cases they can. This is commonly used to improve their performance: https://arxiv.org/abs/2303.11366

load more comments (2 replies)
[–] Asifall@lemmy.world 8 points 1 year ago

I think there’s a big difference between being able to identify an AI by talking to it and being able to identify something written by an AI, especially if a human has looked over it for obvious errors.

[–] Tyler_Zoro@ttrpg.network 3 points 1 year ago (1 children)

What you are describing is true of older LLMs. GPT4, it's less true of. GPT5 or whatever it is they are training now will likely begin to shed these issues.

The shocking thing that we discovered that lead to all of this is that this sort of LLM continues to scale in capabilities with the quality and size of the training set. AI researchers were convinced that this was not possible until GPT proved that it was.

So the idea that you can look at the limitations of the current generation of LLM and make blanket statements about the limitations of all future generations is demonstrably flawed.

[–] jocanib@lemmy.world 2 points 1 year ago (1 children)

They cannot be anything other than stochastic parrots because that is all the technology allows them to be. They are not intelligent, they don't understand the question you ask or the answer they give you, they don't know what truth is let alone how to determine it. They're just good at producing answers that sound like a human might have written them. They're a parlour trick. Hi-tech magic 8balls.

[–] Tyler_Zoro@ttrpg.network 4 points 1 year ago (3 children)

They cannot be anything other than stochastic parrots because that is all the technology allows them to be.

Are you referring to humans or AI? I'm not sure you're wrong about humans...

load more comments (3 replies)
load more comments (2 replies)
[–] Candelestine@lemmy.world 45 points 1 year ago (3 children)

Clearly the Founding Fathers were not advanced enough to have crafted the US Constitution unaided. It's only reasonable to imagine that ancient aliens could have landed, given them an AI to assist them, and then departed with nobody the wiser.

I am certain we can find evidence of this if we dig hard enough.

[–] Addition@sh.itjust.works 16 points 1 year ago

James Madison was a fraud! He used ChatGPT instead of writing out his homework like he was supposed to. Smh my head.

[–] Tyler_Zoro@ttrpg.network 6 points 1 year ago

Clearly the Founding Fathers were not advanced enough to have crafted the US Constitution unaided.

In a sense you are correct. They cribbed from lots of the most well known political philosophers at the time. For example, there are direct quotes from Locke in the Declaration and his influence over the Constitution can be felt clearly.

[–] livus@kbin.social 2 points 1 year ago* (last edited 1 year ago)

I am certain we can find Georgios Tsoukalos saying there is evidence of this sooner or later. Probably extraterrestrial!

[–] dethb0y@lemmy.world 29 points 1 year ago (1 children)

Because AI detectors suck and are the modern day equivalent of dowsing rods?

[–] jocanib@lemmy.world 19 points 1 year ago* (last edited 1 year ago) (1 children)

They're circular. If the text is too predictable it was written by an LLM* but LLMs are designed to regurgitate the next word most commonly used by humans in any given context.

*AI is a complete misnomer for the hi-tech magic 8ball

[–] Zeth0s@lemmy.world 3 points 1 year ago* (last edited 1 year ago) (1 children)

The next most commonly used word would result in a loop of common word. LLMs do not work like that

[–] jocanib@lemmy.world 12 points 1 year ago (2 children)

In context. And that is exactly how they work. It's just a statistical prediction model with billions of parameters.

[–] keegomatic@kbin.social 9 points 1 year ago

That’s not really how LLMs work. You’re basically describing Markov chains. The statement “It’s just a statistical prediction model with billions of parameters” also applies to the human brain. An LLM is much more of a black box than you’re implying.

[–] Zeth0s@lemmy.world 5 points 1 year ago* (last edited 1 year ago) (2 children)

regurgitate the next word most commonly used by humans in any given context.

is not what it does. That would create non sensical text (you can try yourself).

This is a summary of the method, as summarized by gtp-4:


Sure, here is a detailed description of how text is generated with ChatGPT, which is based on the GPT architecture:

  1. Initial Prompt: The process begins with an input prompt. This could be something like "Tell me about the weather today" or any other string of text.
  1. Tokenization: The input text is broken down into smaller parts, called tokens, which can represent words, parts of words, or punctuation. GPT uses a byte pair encoding (BPE) tokenization, which essentially breaks down text into commonly occurring chunks.
  1. Embedding: Each token is then turned into a vector via an embedding. This vector captures semantic information about the token and serves as the input for the model.
  1. Processing the Input: The GPT model processes the input vectors sequentially with a stack of transformer layers. Each layer applies self-attention and feeds its output into the next layer.
  1. Self-Attention Mechanism: The self-attention mechanism in the Transformer model allows it to weigh the importance of different words when predicting the next word. For example, when trying to predict the last word in the sentence "The cat sat on the ____," the words "cat" and "on" are likely to have more influence on the prediction than "The". This weighing is learned during training and allows the model to generate more coherent and contextually appropriate responses.
  1. Output Layer: The output from the final transformer layer for the last input token goes through a linear layer followed by a softmax function, which turns it into a probability distribution over the possible next tokens in the vocabulary. Each possible next token is assigned a probability.
  1. Sampling with Temperature: The next token is chosen based on these probabilities. One common method is to sample from this distribution, which introduces some randomness into the process. The temperature parameter controls the amount of randomness: a higher temperature makes the distribution more uniform and the output more random, while a lower temperature makes the model more likely to choose the highest-probability token.
  1. Decoding: The chosen token is then decoded back into text and appended to the output.
  1. Next Iteration: The process then repeats for the next token: the model takes the output so far (including the newly-generated token), processes it, and generates probabilities for the next token. This continues until a maximum length is reached, or an end-of-sequence token is produced.
  1. Post-Processing: Any necessary post-processing is applied, such as cleaning up tokenization artifacts.

In this way, the model generates a sequence of tokens, one at a time, based on the input prompt and the tokens it has generated so far. Please note that while this process typically uses sampling with a temperature parameter, other methods like beam search or top-k sampling can also be used to choose the next token. These methods have different trade-offs in terms of computational efficiency, diversity, and quality of output.


You are missing the key part where the text is tranformed in a vector space of "concepts" where semanticic relationships are represented, that is where the inference happens. The inference is not on words to get the next commonly used word, otherwise it wouldn't work. And you also missed the final sampling to introduce a randomness in the word selection.

I don't understand why are you so upset for a chain of complex mathematical functions that complete and input sentence. Why are you angry?

load more comments (2 replies)
[–] rarely@sh.itjust.works 27 points 1 year ago (2 children)

As an AI language model, I am unable to found a country but here are some ideas which might help people feel less likely to revolt.

[–] Zeth0s@lemmy.world 6 points 1 year ago

This is a simple example of a constitution. For a real constitution you might want to consider a more detailed list of amendaments and civil rights. Please refer to a professional constitutional writer if you want to create a constitution for productive purposes

[–] Gimly@lemmy.world 4 points 1 year ago (1 children)

GPT-4 is quite willing to help you create the constitution for your own country, well I guess I have to find a name now:

Certainly! Crafting a constitution for a new country is an exciting and monumental task. The constitution is the core document that dictates the legal obligations and rights of your country's government and its citizens. Here's a very basic outline to get us started. Please remember, this is a rudimentary draft and it will need to be adjusted based on the specific needs, culture, history, and goals of your country:


PREAMBLE

We, the people of [Your Country's Name], in order to form a more perfect union, establish justice, ensure domestic tranquility, provide for the common defense, promote the general welfare, and secure the blessings of liberty to ourselves and our posterity, do ordain and establish this Constitution for [Your Country's Name].

ARTICLE I: THE LEGISLATURE

Section 1: All legislative powers herein granted shall be vested in a [name of your legislative body].

Section 2: [Details on how members are elected, their terms, qualifications, powers and responsibilities]

Section 3: [Details on how legislation is passed, veto powers, etc.]

ARTICLE II: THE EXECUTIVE

Section 1: The executive power shall be vested in a [title of your head of state].

Section 2: [Details on how the head of state is elected, their term, qualifications, powers and responsibilities]

Section 3: [Details on how the executive branch operates, and other positions within the executive branch]

ARTICLE III: THE JUDICIARY

Section 1: The judicial power of [Your Country's Name] shall be vested in one Supreme Court, and in such inferior courts as the [Legislative body] may from time to time ordain and establish.

Section 2: [Details on how judges are appointed, their terms, qualifications, powers and responsibilities]

ARTICLE IV: THE RIGHTS OF THE PEOPLE

Section 1: All persons born or naturalized in [Your Country's Name] are citizens of [Your Country's Name] and of the state wherein they reside. No state shall make or enforce any law which shall abridge the privileges or immunities of citizens.

Section 2: [Further details on the rights, freedoms, and responsibilities of the citizens, these could include things like freedom of speech, freedom of religion, right to a fair trial, etc.]

ARTICLE V: AMENDMENTS

Section 1: The [Legislative body] whenever two thirds of both Houses deem it necessary, shall propose Amendments to this Constitution, or, on the Application of the Legislatures of two thirds of the several states, shall call a Convention for proposing Amendments...

ARTICLE VI: GENERAL PROVISIONS

Section 1: [Any other items that need to be addressed, such as the official language, the national religion (or lack thereof), the national emblem, the process for international treaties, etc.]

ARTICLE VII: RATIFICATION

This Constitution shall be considered ratified when agreed upon by a majority of citizens through a free and fair vote.


Please remember that a constitution is a living document that will grow and change over time as your country evolves. This outline is just a starting point and may not include all the specific sections you need. For example, you might want to add sections addressing human rights, environmental protection, financial regulations, and so on.

This process should involve broad input from many different stakeholders, including legal experts, and should be reflective of the country's culture, values, history, and aspirations for the future. Also, please note that the effective implementation of a constitution requires a strong rule of law and mechanisms for holding those in power accountable.

load more comments (1 replies)
[–] Dohnakun@lemmy.fmhy.ml 15 points 1 year ago (1 children)

This article was written to keep people as long on the page as possible. It didn't get to the point before i left. Someone has a tl;dr?

[–] Postcard64@lemmy.world 17 points 1 year ago (2 children)

Constitution is a text that appears many times on the internet. ChatGPT's training set probably has multiple copies of it. So it's likely ChatGPT will generate it. Therefore, the detectors are likely to flag it as AI-generated. That's what I got from it, but I also found it difficult to parse. Maybe someone can correct me on this.

[–] SubArcticTundra@lemmy.ml 8 points 1 year ago (1 children)

I thimk we need and AI to summarize the article. Edit: oh god, soon shit articles are gonna be optimized to not be AI summarizable

[–] Dohnakun@lemmy.fmhy.ml 3 points 1 year ago

Bet it was written by an AI.

load more comments (1 replies)
[–] Kolanaki@yiffit.net 11 points 1 year ago (1 children)

I knew it. The founding fathers were robots from the future!

[–] busturn@lemmy.world 10 points 1 year ago (2 children)

I've recently checked my years-old essay using one of these AI plagiarism detectors and it said that the essay was 90% AI written. So either it's all bs or I'm a time travelling AI.

[–] 98codes@lemm.ee 5 points 1 year ago

I’m convinced that it’s been trained on top of the essays of middle and high school students that have gone their whole lives without proper education on vocabulary, grammar, and the like. So when asked to evaluate something written properly, it’s flagged as AI.

Garbage in, garbage out. Same as is ever was.

load more comments (1 replies)
[–] usrsbin@kbin.social 7 points 1 year ago

Easy fix: run a constitution detector before the AI text detector

[–] Bluefold@sh.itjust.works 5 points 1 year ago

Interesting fact: The same parent company owns both Reddit and Turnitin.

[–] kikuchiyo@lemmy.ml 3 points 1 year ago (1 children)

I’m waiting for new conspiracy theories after that article hahah.

[–] paddirn@lemmy.world 4 points 1 year ago

Obviously the US Constitution was written by AI, we're living in a simulation. Wake up sheeple, the Matrix is real!

[–] skellener@kbin.social 3 points 1 year ago

Don’t mistake artificial intelligence for actual intelligence.

[–] Captain_Patchy@lemmy.world 2 points 1 year ago

They only know what they have been fed.

What more likely first/base feeding than the US Constitution's declarations and it's amendments?

[–] rarely@sh.itjust.works 2 points 1 year ago

This. Changes. Everything.

[–] someguy3@lemmy.ca 2 points 1 year ago (1 children)
load more comments (1 replies)
load more comments
view more: next ›