this post was submitted on 11 Aug 2023
53 points (81.2% liked)
Programming.dev Meta
2466 readers
1 users here now
Welcome to the Programming.Dev meta community!
This is a community for discussing things about programming.dev itself. Things like announcements, site help posts, site questions, etc. are all welcome here.
Links
Credits
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
My experience has often been the opposite. Programmers will do a lot to avoid the ethical implications of their works being used maliciously and discussions of what responsibility we bear for how our work gets used and how much effort we should be obligated to make towards defending against malicious use.
It's why I kind of wish that "engineer" was a regulated title in America like it is in other countries, and getting certified as a programming engineer required some amount of training in programming ethics and standards.
How would a formal licensing system work for software engineering? How would they keep up with the rapid evolution in this industry?
I believe in better education in this field, but the standard "engineer" programs from other fields don't translate to software. Having the government codify today's standards would stunt the industry as a whole and kill innovation. Imagine if they had done that in the 90s and said all programming must be waterfall, monolithic, relational dbs, and using c/Fortran/Cobol.
Maybe I just don't understand how other countries handle it though. I know my country would absolutely screw it up
Formal licensing could be about things that are language agnostic. How to properly use tests to guard against regressions, how to handle error states safely.
How do you design programs for critical systems that CANNOT fail, like pace makers? How do you guard against crashes? What sort of redundancy do you need in your software?
How do you best design error messages to tell an operator how to fix the issue? Especially in critical systems like a plane, how do you guard against that operator doing the wrong thing? I'm thinking of the DreamLiner incidents where the pilots' natural inclination was to grab the yoke and pull up, which unknowingly fought the autopilot and caused the plane to stall. My understanding was that the error message that triggered during those crashes was also extremely opaque and added further confusion in a life-and-death situation.
When do you have an ethical responsibility not to ship code? Just for physical safety? What about Dark Patterns? How do you recognize them and do you have an ethical responsibility to refuse implementation? Should your accreditation as an engineer rely on that refusal, giving you systemic external support when you do so?
None of that is impacted by what tech stack you are using. They all come down to generic logical and ethical reasoning.
Lastly, under certain circumstances, Civil engineers can be held personally liable for negligence when their bridge fails and people die. If we are going to call ourselves "engineers", we should bear the same responsibility. Obviously not every software developer needs to have such high standards, but that's why software engineer should mean something.