For Artificial Humor we'd need to add a confidence matrix of expectations. - with different layers of certainties. Humor isn't hard to decode. It's changing context on several levels simultaneously. Some we expect, which makes it funny (fits pattern-of-funny, and the laugh typically comes after a certain set of parameters are met, which requires timing) and then there is the unexpected change of context, where we are in a safe environment and the context / world frame change isn't deadly. Safe/unsafe - that's another level that needs to be added to the confidence. In an unsafe environment, unless one has cognitive differences, "THIS ISN'T FUNNY!" is a common response; yet then so is laughter... it "lightens the mood" because one laughs at the cognitive dissonance to help reconcile it (laughing your way into acceptance) - turning the XOR into an == as it were. Anyway, I can visualize the humor program sending and receiving in pseudo-code in my head but I'm sure somebody else has drawn it out, labelled it, has humor programs that can comprehend jokes already. If they don't, they will, because if I can see it, I'm sure somebody else 'out there' has too, or will