Formerly /u/Zalack on Reddit.e

Also [email protected]

  • 0 Posts
  • 26 Comments
Joined 11 months ago
cake
Cake day: August 3rd, 2023

help-circle









  • Formal licensing could be about things that are language agnostic. How to properly use tests to guard against regressions, how to handle error states safely.

    How do you design programs for critical systems that CANNOT fail, like pace makers? How do you guard against crashes? What sort of redundancy do you need in your software?

    How do you best design error messages to tell an operator how to fix the issue? Especially in critical systems like a plane, how do you guard against that operator doing the wrong thing? I’m thinking of the DreamLiner incidents where the pilots’ natural inclination was to grab the yoke and pull up, which unknowingly fought the autopilot and caused the plane to stall. My understanding was that the error message that triggered during those crashes was also extremely opaque and added further confusion in a life-and-death situation.

    When do you have an ethical responsibility not to ship code? Just for physical safety? What about Dark Patterns? How do you recognize them and do you have an ethical responsibility to refuse implementation? Should your accreditation as an engineer rely on that refusal, giving you systemic external support when you do so?

    None of that is impacted by what tech stack you are using. They all come down to generic logical and ethical reasoning.

    Lastly, under certain circumstances, Civil engineers can be held personally liable for negligence when their bridge fails and people die. If we are going to call ourselves “engineers”, we should bear the same responsibility. Obviously not every software developer needs to have such high standards, but that’s why software engineer should mean something.


  • My experience has often been the opposite. Programmers will do a lot to avoid the ethical implications of their works being used maliciously and discussions of what responsibility we bear for how our work gets used and how much effort we should be obligated to make towards defending against malicious use.

    It’s why I kind of wish that “engineer” was a regulated title in America like it is in other countries, and getting certified as a programming engineer required some amount of training in programming ethics and standards.









  • I agree with the other poster that you need to define what you even mean when you say free will. IMO, strict determinism is not incompatible with free will. It only provides the mechanism. I posted this in another thread where this came up:

    The implications of quantum mechanics just reframes what it means to not have free will.

    In classical physics, given the exact same setup you make the exact same choice every time.

    In Quantum mechanics, given the same exact setup, you make the same choice some percentage of the time.

    One is you being an automaton while the other is you being a flipped coin. Neither of those really feel like free will.

    Except.

    We are looking at this through an implied assumption that the brain is some mechanism, separate from “us”, which we are forced to think “through”. That the mechanisms of the brain are somehow distorting or restricting what the underlying self can do.

    But there is no deeper “self”. We are the brain. We are the chemical cascade bouncing around through the neurons. We are the kinetic billiard balls of classical physics and the probability curves of quantum mechanics. It doesn’t matter if the universe is deterministic and we would always have the same response to the same input or if it’s statistical and we just have a baked “likelihood” of that response.

    The way we respond or the biases that inform that likelihood is still us making a choice, because we are that underlying mechanism. Whether it’s deterministic or not it’s just an implementation detail of free will, not a counterargument.