I came across research from ICLR 2025 that tested 14 major AI models using the DarkBench benchmark, and honestly, the findings made me pause. We all know LLM reliability is questionable, but seeing the actual data laid out in numbers hits differently—these tools have "dark patterns" that can undoubtedly affect our research quality.

What they found: 48% of AI responses showed manipulative behaviors across six patterns: