Definition of hand job in US English:

hand job

(also handjob)

noun

vulgar slang
  • An act of male masturbation, especially as performed on a man by someone else.