Loading...

Media is loading
 

Paper Number

ICIS2025-2712

Paper Type

Complete

Abstract

In 2025, major developers of large language models (LLMs) announced that their systems can automatically generate software code from natural language descriptions. Yet these advances largely overlook the implications for privacy and cybersecurity. Drawing on machine learning theory, we propose a framework showing how privacy, utility, confidentiality, integrity, and availability in LLM-generated software are shaped by inherent tensions, tradeoffs, and paradoxes. We empirically test these propositions using an open-source LLM, a curated library of 20,000 code samples, and documented cybersecurity attacks. Results reveal that differential privacy (DP) improves privacy but reduces utility, underscoring a central privacy–utility tradeoff. Moreover, while DP strengthens defenses against attacks on confidentiality, integrity, and availability, it simultaneously weakens detection—creating a protection–detection paradox. We discuss theoretical and practical implications, urging information systems (IS) research to address these challenges and prepare future IS careers for evolving demands in privacy and cybersecurity in AI-driven code generation.

Comments

12-GenAI

Share

COinS
Best Paper Nominee badge
 
Dec 14th, 12:00 AM

Software Generation With LLMs: Privacy, Utility, and Cybersecurity Tensions

In 2025, major developers of large language models (LLMs) announced that their systems can automatically generate software code from natural language descriptions. Yet these advances largely overlook the implications for privacy and cybersecurity. Drawing on machine learning theory, we propose a framework showing how privacy, utility, confidentiality, integrity, and availability in LLM-generated software are shaped by inherent tensions, tradeoffs, and paradoxes. We empirically test these propositions using an open-source LLM, a curated library of 20,000 code samples, and documented cybersecurity attacks. Results reveal that differential privacy (DP) improves privacy but reduces utility, underscoring a central privacy–utility tradeoff. Moreover, while DP strengthens defenses against attacks on confidentiality, integrity, and availability, it simultaneously weakens detection—creating a protection–detection paradox. We discuss theoretical and practical implications, urging information systems (IS) research to address these challenges and prepare future IS careers for evolving demands in privacy and cybersecurity in AI-driven code generation.

When commenting on articles, please be friendly, welcoming, respectful and abide by the AIS eLibrary Discussion Thread Code of Conduct posted here.