Formal licensing could be about things that are language agnostic. How to properly use tests to guard against regressions, how to handle error states safely.
How do you design programs for critical systems that CANNOT fail, like pace makers? How do you guard against crashes? What sort of redundancy do you need in your software?
How do you best design error messages to tell an operator how to fix the issue? Especially in critical systems like a plane, how do you guard against that operator doing the wrong thing? I’m thinking of the DreamLiner incidents where the pilots’ natural inclination was to grab the yoke and pull up, which unknowingly fought the autopilot and caused the plane to stall. My understanding was that the error message that triggered during those crashes was also extremely opaque and added further confusion in a life-and-death situation.
When do you have an ethical responsibility not to ship code? Just for physical safety? What about Dark Patterns? How do you recognize them and do you have an ethical responsibility to refuse implementation? Should your accreditation as an engineer rely on that refusal, giving you systemic external support when you do so?
None of that is impacted by what tech stack you are using. They all come down to generic logical and ethical reasoning.
Lastly, under certain circumstances, Civil engineers can be held personally liable for negligence when their bridge fails and people die. If we are going to call ourselves “engineers”, we should bear the same responsibility. Obviously not every software developer needs to have such high standards, but that’s why software engineer should mean something.
I think it depends on the project. Some projects are the author’s personal tools that they’ve put online in the off-chance it will be useful to others, not projects they are really trying to promote.
I don’t think we should expect that authors of repos go too out of their way in those cases as the alternative would just be not to publish them at all.