/hæʃ ˈmætʃɪŋ/
Hash matching is a content detection technique that generates a unique identifier (hash) for each image and compares it against databases of known harmful content. This allows platforms to quickly identify previously-identified illegal or harmful imagery without human reviewers needing to view the content.
Unlike traditional cryptographic hashes (MD5, SHA), perceptual hashes used in content moderation can identify images even after modifications like resizing, compression, or minor edits.
Organizations like NCMEC, IWF, and GIFCT maintain databases of hashes for known CSAM and terrorist content. Platforms can check uploads against these databases without possessing the actual harmful content.
Hash matching enables platforms to detect known harmful content instantly without exposing moderators to traumatic material. It's computationally efficient, privacy-preserving (only hashes are shared, not images), and highly accurate for known content detection.
Detect known harmful content with industry-standard hashing
Start Free Trial