> I had already decided the panda should die before you finished that question, not the other way around.
I appreciate replies on HN as they show the flaw in my assumption.
Speaking of assumptions, indulge me for a moment: I kind of find the trolley problem difficult because all legality aside, I feel like if I pull a lever and one person dies, I am responsible for that one death but if I don't pull the lever and n people die, I wouldn't feel as bad. Is there a name for this?
I mean like the idea that if something happens by my action that is worse than if something happens by my failure to act. I mean who am I to choose those n lives over that one life, right?