AI Voice Cloning

The availability of a few seconds of an audio sample (samples of even less than five seconds) is enough to train AI to replicate a person’s voice resembling the accent, pauses, speech patterns, etc. After this, a text to speech software is used. All the inputs will be cloned.

Cloning the voice of celebrities after their death or giving an audio outlet to people who are unable to speak could be the legitimate uses of voice cloning. However, such cloning could be used for criminal purposes. It is easier to get the voice samples. It could be an answer to a wrong number. Or a video uploaded by a person about his life and work. Or a public speech or an interview. Or hosting a programme.

The cloned voice could be robotic. It could sound like gibberish. The caller could be in a hurry. It is not wise to reveal confidential data. One can be alert and can cross-examine the caller. However, the voice at the other end sounds so convincing.

Voice cloning is a serious risk.

print

Comments

Leave a Reply

Your email address will not be published. Required fields are marked *