A new survey from Legit Security uncovers significant consumer apprehension toward AI-generated code in applications, with nearly half expressing worries about potential vulnerabilities. Conducted among 1,000 U.S. consumers, the findings highlight the critical need for transparency and robust security practices as AI integrates deeper into software development.
The survey, commissioned by Legit Security and executed by Dynata, reveals that while AI adoption in software is inevitable, consumer trust hinges on responsible implementation. One in four respondents would abandon their preferred applications upon discovering AI-written code, emphasizing the erosion of confidence when vulnerabilities arise from AI processes.
Security vulnerabilities top consumer fears at 34%, followed by unpredictable application behavior at 23% and data training issues at 21%. Factors boosting perceived security include official app stores at 53%, privacy policies at 46%, and well-known brands at 45%. These insights stress the importance of visible accountability in AI-driven development to maintain user confidence.
The report identifies stark generational differences in AI tolerance. Over 40% of Boomers fear AI vulnerabilities and are twice as likely to lose trust upon AI disclosure. In contrast, younger users demonstrate resilience, with Gen Z showing increased trust in AI-enhanced applications for 34%, reflecting varied attitudes toward technological innovation.
"AI itself isn't a dirty word to consumers. The real issue is whether companies use it responsibly," says Roni Fuchs, co-founder and CEO at Legit. "Most people don't reject apps just because they leverage AI-generated code. Many of them understand it's inevitable. The real breaking point comes when AI introduces a vulnerability. At that moment, trust erodes fast, and potentially permanently. As AI adoption accelerates across the software development lifecycle, the mandate is clear: companies must make preventing, detecting, and remediating vulnerabilities in AI-generated code a non-negotiable priority. Anything less risks losing your users' trust."
"There is urgency for engineering teams to ensure that AI-generated code can be safe, secure, and trustworthy," says Liav Caspi, co-founder and CTO at Legit. "Cybersecurity Awareness Month has traditionally emphasized consumer best practices, but these findings highlight that developer practices matter a lot, too. Users are right to be concerned about how AI is being leveraged in the applications they use daily, and while they will keep downloading apps with AI, visible signals of security and accountability are needed to create this trust."
Released at the onset of National Cybersecurity Awareness Month, the survey calls for engineering teams to prioritize secure AI code practices. Explore the full findings in Legit Security's blog post and visit www.legitsecurity.com for solutions in AI-native application security.
The Legit Security AI-Native ASPM platform is a new way to manage application security in a world of AI-first development, providing a cleaner way to manage and scale AppSec and address risks. Fast to implement, easy to use, and AI-native, Legit has an unmatched ability to discover and visualize the entire software factory attack surface, including a prioritized view of AppSec data from siloed scanning tools. As a result, organizations have the visibility, context, and automation they need to quickly find, fix, and prevent the application risk that matters most. Spend less time chasing low-risk findings and more time innovating.