APA (7th ed.) Citation

Omar, M., Sorin, V., Collins, J. D., Reich, D., Freeman, R., Gavin, N., . . . Klang, E. Multi-model assurance analysis showing large language models are highly vulnerable to adversarial hallucination attacks during clinical decision support. Nature Portfolio.

Chicago Style (17th ed.) Citation

Omar, Mahmud, et al. Multi-model Assurance Analysis Showing Large Language Models Are Highly Vulnerable to Adversarial Hallucination Attacks During Clinical Decision Support. Nature Portfolio.

MLA (9th ed.) Citation

Omar, Mahmud, et al. Multi-model Assurance Analysis Showing Large Language Models Are Highly Vulnerable to Adversarial Hallucination Attacks During Clinical Decision Support. Nature Portfolio.

Warning: These citations may not always be 100% accurate.