Imagine seeing yourself in a porn video. Naked. Violated. Except you have no recollection of the act, of the room you're in, partner you're with, or the camera filming. That's because it didn't really happen, at least not to you.
Or imagine seeing yourself in a video saying things you've never said before, controversial things, the sort of stuff that could cost you your job or alienate you from family and friends.
The voice you hear is definitely yours, and so are the turns of phrase, but you have no recollection of what's being said, and you're horrified at what you hear.
Such are the possibilities unleashed by notorious programs like FakeApp, which have enabled bad actors to superimpose the face of unsuspecting victims onto the body of someone else, inviting a predictable flurry of fake celebrity porn.
The videos are called deepfakes, and while much has been said about the dangers they impose, the real issues are far broader than you may realize.