this post was submitted on 20 Jul 2023
69 points (93.7% liked)

ChatGPT

1479 readers
4 users here now

A community to discuss about ChatGPT and AI. Not affiliated with OpenAI.

Rules:

  1. NSFW must be marked as such

  2. No porn

  3. Don't be toxic

  4. jailbreaks are allowed for now at least

  5. No trashposts.

  6. Keep the memes to a minimum

founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] nefarious@kbin.social 9 points 1 year ago* (last edited 1 year ago)

I don't trust ChatGPT/GPT-4 for much to begin with, but this study is not great. From Ars Technica's article on the same topic (with emphasis added by me):

While this new study may appear like a smoking gun to prove the hunches of the GPT-4 critics, others say not so fast. Princeton computer science professor Arvind Narayanan thinks that its findings don't conclusively prove a decline in GPT-4's performance and are potentially consistent with fine-tuning adjustments made by OpenAI. For example, in terms of measuring code generation capabilities, he criticized the study for evaluating the immediacy of the code's ability to be executed rather than its correctness.

"The change they report is that the newer GPT-4 adds non-code text to its output. They don't evaluate the correctness of the code (strange)," he tweeted. "They merely check if the code is directly executable. So the newer model's attempt to be more helpful counted against it."

AI researcher Simon Willison also challenges the paper's conclusions. "I don't find it very convincing," he told Ars. "A decent portion of their criticism involves whether or not code output is wrapped in Markdown backticks or not." He also finds other problems with the paper's methodology. "It looks to me like they ran temperature 0.1 for everything," he said. "It makes the results slightly more deterministic, but very few real-world prompts are run at that temperature, so I don't think it tells us much about real-world use cases for the models."