In the age of replicated AI voices and security it seems like this would be a bad idea. How much voice sample is needed to decently replicate someone's voice?
Think of the old scam where someone texts you and says they're your granddaughter and they're stuck somewhere and need an urgent money transfer. In the old days the advice would be: call your granddaughter and have get confirm the story. But now with AI tech your granddaughter could call you up and deliver the scam in her own voice.