• oshitwaddup@lemmy.antemeridiem.xyz
    link
    fedilink
    arrow-up
    1
    ·
    10 months ago

    i think I do understand them, I’ve thought about that problem before. Can you go into more detail on what you mean by internally inconsistent? By my understanding, situations in the world can come about where values need to be weighed, or there are only bad choices available, but that doesn’t mean those values should be discarded or replaced or that they shouldn’t be shared/spread.

    • commie@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      1
      ·
      10 months ago

      either it’s true that you can write an axiom that says “sentient beings should always consent to anything that is done to them” or you can write an axiom that says “you should always do what will bring about the most happiness or at least distress”

      those axioms are in conflict with one another. it’s not that there’s only bad choices. it’s that you’ve given yourself conflicting standards.

      • oshitwaddup@lemmy.antemeridiem.xyz
        link
        fedilink
        arrow-up
        1
        ·
        10 months ago

        Neither of those are axioms I hold. The axiom “all sentient beings are morally relevant” does not specify how to go from there, and I am not convinced that any one ethical framework is “the one”. There are some things that all the ones I’m aware of converge on with a sentientist perspective, but there are weird cases as well like whether to euthanize stray animals where they don’t converge