Next Article in Journal
Electromagnetic Interference Simulation and Shielding Design for Aircraft Engine Nacelle Subjected to EMALS
Previous Article in Journal
Development of an Embedded In-Mass Inertial Device for Landslide and Rockfall Monitoring
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Article

Levels of Automated Code Generation (LACG): An Operational Taxonomy for AI-Augmented Software Construction

1
College of Computer and Information Sciences, Fujian Agriculture and Forestry University, Fuzhou 350002, China
2
The Fujian Provincial Key Laboratory of Smart Agriculture and Forestry, Fujian Agriculture and Forestry University, Fuzhou 350002, China
3
Fujian Key Laboratory of Severe Weather, Fuzhou 350002, China
4
Key Laboratory of Straits Severe Weather, China Meteorological Administration, Fuzhou 350002, China
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2026, 16(10), 4788; https://doi.org/10.3390/app16104788
Submission received: 16 April 2026 / Revised: 6 May 2026 / Accepted: 7 May 2026 / Published: 11 May 2026
(This article belongs to the Topic Applications of NLP, AI, and ML in Software Engineering)

Abstract

Artificial intelligence (AI) coding systems now range from inline completion to repository-level agents and platform-supported application builders; yet, software engineering still lacks a code-generation-centered operational taxonomy for describing how much work is delegated, under what conditions, and with what responsibility structure. This study proposes Levels of Automated Code Generation (LACG), a six-level taxonomy (L0–L5) for classifying automation in AI-augmented software construction. LACG is organized around four responsibility-aware concepts—The Software Development Task (SDT), Operational Capability Domain (OCD), fallback responsibility, and minimal risk condition—and is assigned to a declared configuration–SDT–OCD tuple rather than to a vendor brand or model family in the abstract. To reduce the risk that public vendor documentation reproduces marketing bias, the method separates declared affordance evidence from routine capability evidence and adopts an evidence-triangulation design. Public documentation is used only to identify configuration boundaries and declared affordances; independent software engineering benchmarks, agent studies, productivity studies, and taxonomy-evaluation literature are used to calibrate the level boundaries and constrain the claims. LACG is then applied to 30 representative current AI coding tool configurations using time-stamped public-documentation records, with boundary logic cross-checked against independent evidence on repository-level issue solving, agent tool use, and context-dependent productivity outcomes. Three anonymized human raters, selected for software engineering or AI-coding-tool expertise and independent of the authors and evaluated vendors, then classified the same prepared, blinded public-documentation records using the LACG coding manual. Exact three-rater agreement was 28/30 (93.3%); adjacent-level and majority agreement were both 30/30 (100.0%); mean pairwise quadratic-weighted Cohen’s kappa was 0.963; and Krippendorff’s alpha for ordinal ratings was 0.963. These agreement statistics test classification consistency over a structured documentary evidence base; they do not test actual tool behavior, direct execution, product performance, safety, productivity, or deployment outcomes. After adjudication, the final sample contains six L1 configurations, nine L2 configurations, and fifteen L3 configurations; no public configuration is classified as L4 or L5 under the fallback-responsibility criterion. The study supports preliminary, documentation-bound classification applicability, boundary calibration, and discriminative vocabulary development, not predictive validation or product-level performance claims. LACG provides an operational vocabulary for future empirical work on AI-augmented software construction, benchmark design, tool comparison, and responsibility allocation, while leaving outcome validation for governance, security, productivity, and procurement to subsequent empirical studies.
Keywords: AI-intensive software systems; AI-augmented software construction; software quality and trustworthiness; taxonomy; utility demonstration; code generation AI-intensive software systems; AI-augmented software construction; software quality and trustworthiness; taxonomy; utility demonstration; code generation

Share and Cite

MDPI and ACS Style

Chen, Z.; Lin, L.; Lin, X.; Chen, Y.; Wang, L. Levels of Automated Code Generation (LACG): An Operational Taxonomy for AI-Augmented Software Construction. Appl. Sci. 2026, 16, 4788. https://doi.org/10.3390/app16104788

AMA Style

Chen Z, Lin L, Lin X, Chen Y, Wang L. Levels of Automated Code Generation (LACG): An Operational Taxonomy for AI-Augmented Software Construction. Applied Sciences. 2026; 16(10):4788. https://doi.org/10.3390/app16104788

Chicago/Turabian Style

Chen, Zhenhan, Lizheng Lin, Xiaoyu Lin, YingXin Chen, and Lijin Wang. 2026. "Levels of Automated Code Generation (LACG): An Operational Taxonomy for AI-Augmented Software Construction" Applied Sciences 16, no. 10: 4788. https://doi.org/10.3390/app16104788

APA Style

Chen, Z., Lin, L., Lin, X., Chen, Y., & Wang, L. (2026). Levels of Automated Code Generation (LACG): An Operational Taxonomy for AI-Augmented Software Construction. Applied Sciences, 16(10), 4788. https://doi.org/10.3390/app16104788

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop