Unmasking Vulnerabilities: Adversarial Attacks via Word-Level Manipulation on NLP Models