Learning from human preference