this post was submitted on 03 Nov 2023
30 points (65.3% liked)

Programming

17511 readers
301 users here now

Welcome to the main community in programming.dev! Feel free to post anything relating to programming here!

Cross posting is strongly encouraged in the instance. If you feel your post or another person's post makes sense in another community cross post into it.

Hope you enjoy the instance!

Rules

Rules

  • Follow the programming.dev instance rules
  • Keep content related to programming in some way
  • If you're posting long videos try to add in some form of tldr for those who don't want to watch videos

Wormhole

Follow the wormhole through a path of communities !webdev@programming.dev



founded 1 year ago
MODERATORS
 

It writes more informative commits than I could ever make so I'm just reading what it says and mostly copy/pasting completely most of the time, I write all of the changes I've made into an LLM with a large context window and it write a very detailed commit not just with a title but with bullet points describing each of the changes precisely

you are viewing a single comment's thread
view the rest of the comments
[โ€“] TootSweet@lemmy.world 11 points 1 year ago* (last edited 1 year ago) (1 children)

If you were on my team and I knew you were doing this, aside from all the other issues everybody else has mentioned in this thread, I'd be going out of my way not just to check every single one of your git commits (both code changes and commit messages) for inaccuracies but also also to find every possible reason to nit pick everything you committed.

You shouldn't be using LLMs to write your git commits (messages or code changes.) They hallucinate. But if you are going to use them, you need to spend much more time proofreading what they output than you'd spend writing it yourself. Check every single word for errors. (And, honestly, make the text fit with the way other commits by your team read (assuming you are on a team).)

In short, if you're going to use LLMs, DO NOT TRUST ANYTHING THEY GIVE YOU. And don't be surprised if you get negative blowback from others for using them at all. Keep in mind what can happen if you trust LLMs.

[โ€“] emma@programming.dev 3 points 1 year ago* (last edited 1 year ago)

IMHO, the provided link is largely irrelevant to this topic. It is about lawyers who used ChatGPT as a search engine https://youtu.be/oqSYljRYDEM?t=1436, which is not what it is for, and it will tell you that over and over again. The lawyers in question were not even "trusting ChatGPT". They blatantly and actively disregarded ChatGPT telling them that it was not a search engine and could not provide legal advice https://youtu.be/oqSYljRYDEM?t=1466

This topic is about using LLMs to generate natural language describing code changes that it is provided with which is not only completely different than using an LLM as a defacto search engine, but it is also something LLMs are actually meant to do: autocomplete. This topic is more akin to using LLMs to write title headings for legal documents which are already basically complete as is than it is akin to the link provided.