FBI advises smartphone users to hang up on scammers
The FBI is warning smartphone users about a rise in AI-powered attacks targeting iPhone and Android devices. Security experts say that these attacks include deepfake technology that can mimic voices, making it easier for scammers to trick victims. The dangers of these scams have increased as voice cloning tools have become cheaper and more effective. Cybersecurity expert Adrianus Warmenhoven highlighted that criminals often impersonate family members in emergency situations, requesting money. These tactics can be very convincing, leading many people to fall victim to fraud. To protect yourself, the FBI advises users to hang up immediately if they receive unexpected calls asking for money. Instead, verify the caller's identity through a different method. Both the FBI and Warmenhoven recommend creating a secret code or phrase known only to close friends and family. This code can help confirm someone's identity during a call that might seem suspicious. Deepfake technology relies on audio clips from social media and other sources to replicate a person’s voice. As these attacks become more common and sophisticated, it’s essential to stay cautious and informed.