Next Article in Journal
Developing a Crowdsourcing Digital Repository for Natural and Cultural Heritage Preservation and Promotion: A Report on the Experience in Zakynthos Island (Greece)
Next Article in Special Issue
Cross-Dataset Data Augmentation Using UMAP for Deep Learning-Based Wind Speed Prediction
Previous Article in Journal
Evaluating Optimal Deep Learning Models for Freshness Assessment of Silver Barb Through Technique for Order Preference by Similarity to Ideal Solution with Linear Programming
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

EnterpriseAI: A Transformer-Based Framework for Cost Optimization and Process Enhancement in Enterprise Systems

by
Shinoy Vengaramkode Bhaskaran
Zoom Video Communications, Sanjose, CA 95113, USA
Computers 2025, 14(3), 106; https://doi.org/10.3390/computers14030106
Submission received: 27 January 2025 / Revised: 18 February 2025 / Accepted: 10 March 2025 / Published: 16 March 2025
(This article belongs to the Special Issue Machine Learning and Statistical Learning with Applications 2025)

Abstract

:
Coordination among multiple interdependent processes and stakeholders and the allocation of optimal resources make enterprise systems management a challenging process. Even for experienced professionals, it is not uncommon to cause inefficiencies and escalate operational costs. This paper introduces EnterpriseAI, a novel transformer-based framework designed to automate enterprise system management. This transformer model has been designed and customized to reduce manual effort, minimize errors, and enhance resource allocation. Moreover, it assists in decision making by incorporating all interdependent and independent variables associated with a matter. All of these together lead to significant cost savings across organizational workflows. A unique dataset has been derived in this study from real-world enterprise scenarios. Using the transfer learning approach, the EnterpriseAI transformer has been trained to analyze complex operational dependencies and deliver context-aware solutions related to enterprise systems. The experimental results demonstrate EnterpriseAI’s effectiveness, achieving an accuracy of 92.1%, a precision of 92.5%, and a recall of 91.8%, with a perplexity score of 14. These results represent the ability of the EnterpriseAI to accurately respond to queries. The scalability and resource utilization tests reflect the astonishing factors that significantly reduce resource consumption while adapting to demand. Most importantly, it reduces the operational cost while enhancing the operational flow of business.

1. Introduction

Modern large-scale organizations use enterprise systems as their operational backbone. Enterprise systems integrate various processes, resources, and stakeholders to achieve the objectives of the organization [1]. However, managing these complex systems poses challenges, including inefficiencies in workflow coordination, resource allocation, and decision making [2]. The most common way to respond to these challenges is through manual intervention supported by data analysis and visualization tools. However, this introduces additional delays and the potential occurrence of human errors. Both delays and errors increase operational costs [3].
The integration of AI technologies into enterprise systems offers a promising solution to these challenges. AI models, particularly transformer-based large language models (LLMs), possess unique capabilities that align with the needs of enterprise environments. These models can process vast amounts of both structured and unstructured operational data, identify hidden patterns, and automate decision-making processes. Such automation significantly reduces reliance on human intervention, minimizing errors and improving efficiency. Furthermore, AI systems can provide contextual insights by analyzing interdependencies across different enterprise processes, enabling more informed resource allocation and enhancing overall organizational performance. The proposed EnterpriseAI framework harnesses these AI capabilities, specifically leveraging a transformer model, to address the operational complexities inherent in enterprise systems.
Recent advancements in AI have shown tremendous potential in automating and optimizing system management processes [4]. Significantly, transformer-based large language models have demonstrated their ability to process and analyze vast amounts of data and make decisions from them [5]. These can process both structured and unstructured data and draw conclusions with human-like reasoning [6]. That is why it these models provide promising technologies for integration with enterprise systems, replacing the manual effort of covering operational challenges. This paper introduces EnterpriseAI, a transformer-based framework specifically designed to address the complexities of enterprise systems. The concepts of NLP and transfer learning have been used in this study to automate routine tasks, minimize manual errors, and optimize resource utilization in enterprise systems. As a result, a substantial amount of expenditure can be saved, and operational efficiency can be improved.
The novelty of this study lies in the successful integration of a transformer model trained on a customized dataset derived from enterprise-specific scenarios. The framework effectively uses the model and optimizes the workflow, performing resource planning and mitigating risks. The model’s architecture enables it to understand and analyze the interdependencies among various components of an enterprise system, delivering contextually relevant insights and solutions. The core contributions of this paper are listed below:
  • Technology adaptation: The potential of the transformer model has been utilized in the enterprise system and manual efforts have been effectively automated.
  • Cost optimization: Significant reductions in operational costs have been achieved through automated and accurate system management.
  • Feasibility analysis: A feasibility analysis has been performed through a scalability test and resource utilization observation for real-world enterprise systems integration.
The remainder of this paper is organized as follows. Section 2 provides a comprehensive review of related work. Section 3 outlines the methodology and conceptual framework of EnterpriseAI. Section 4 presents the use case integration. The implementation details are discussed in Section 5, followed by performance evaluation and results in Section 6. The limitations and potential directions for future work are highlighted in Section 7, and Section 8 concludes the paper with key insights.

2. Literature Review

A systematic literature review (SLR) has been conducted to identify research gaps and specify the corresponding research objectives upon exploring the current advancements and challenges in enterprise system management and the adaptation of AI in this field. The guidelines established by Kitchenham et al. [6] have been followed in this study to ensure a comprehensive and structured approach to analyzing the existing literature.

2.1. Research Questions

The literature review has been developed through four research questions (RQs). These questions are directly related to the objectives of this study. The literature review was guided by the following research questions:
  • RQ1: What are the existing challenges in managing enterprise systems?
  • RQ2: How is AI being utilized to address these challenges in enterprise systems?
  • RQ3: What are the limitations of current AI-based approaches in enterprise systems?
  • RQ4: How can transformer-based models enhance the management of enterprise systems?

2.2. Search Strategy

Reliable and authentic sources have been used in this study to identify relevant papers. These sources are IEEE Xplore, SpringerLink, ACM Digital Library, and Scopus [7]. With the aid of the Google Scholar search engine, using search key-phrases to explore these sources only, relevant papers were identified [8]. The search terms were combined as follows:
“enterprise systems” AND “artificial intelligence” AND “transformer models” AND “process optimization” AND “cost reduction”.
Additional filters were applied to include peer-reviewed articles published between 2020 and 2025. However, a few exceptions were made for globally accepted theories, which have been cited multiple times. For these types of papers, the publication time range was relaxed.

2.3. Inclusion and Exclusion Criteria

Not every paper retrieved through the search strategy explained earlier is appropriate for the study conducted in this paper. That is why, even after using a well-developed search strategy, additional inclusion and exclusion criteria have been specified to enhance the quality of the literature review further and ensure relevancy. Both the inclusion and exclusion criteria are presented below:
  • Inclusion Criteria:
    Studies addressing enterprise systems management.
    Research involving the application of AI in enterprise systems.
    Papers discussing transformer-based models or NLP in enterprise contexts.
    Peer-reviewed articles published in reputable journals or conferences.
  • Exclusion Criteria:
    Non-English-language studies.
    Studies lacking experimental results or practical implementations.
    Duplicate publications or outdated reviews.
The selected studies were analyzed to extract key information, including the challenges addressed, AI techniques employed, and reported outcomes. A total of 72 articles were reviewed after applying the inclusion and exclusion criteria. The findings were synthesized and categorized based on the research questions.

2.4. Findings

2.4.1. Challenges in Enterprise Systems

Jiang et al. [9] explored the challenges of enterprise systems from a cybersecurity perspective. A similar study was conducted by Shi and Jincheng [10], which identified the challenges of enterprise systems from a business perspective. According to these studies, inefficient resource allocation, lack of real-time adaptability, and high operational costs are the major challenges facing these systems. These challenges are exacerbated by the complexity of coordinating multiple workflows, stakeholders, and technologies [11].

2.4.2. AI Applications in Enterprise Systems

Application of AI in enterprise systems is not a new concept [12]. Industry 4.0 technologies [13], utilization of Internet of Things (IoT) data [14], and automatic resource planning [15] are a few of the applications of AI in enterprise systems. A comprehensive review conducted by Rehman et al. [16] shows the challenges of using machine learning and IoT in enterprise systems. It summarized a set of solutions as well. Jawad et al. [17] studied various optimization techniques for enterprise resource planning (ERP) systems using machine learning. The existing literature review indicates the growing adoption of AI in enterprise systems. However, the application of AI is confined to non-generative AI approaches, where the AI module is dedicated to making predictions about certain variables.

2.4.3. Limitations of Current AI Approaches

The primary limitation of the current AI approaches in the enterprise system is the dependence on discriminative and analytical AI [18]. The existing AI solutions are focused on making predictions, performing classification, or making small-scale decisions based on pattern recognition [19]. The adaptation of generative AI-based solutions is underexplored. At the same time, technologies presented in the recent literature related to AI are concentrated on the development of the AI model instead of presenting feasible ways of incorporating them into enterprise systems. As a result, most of these solutions seem promising as potential solutions, whereas the real-world feasibility is yet to be explored [20].

2.4.4. Potential of Transformer Models in Enterprise Systems

Transformer models have revolutionized the NLP sector. The large language models (LLMs) developed using transformers as the core technologies are capable of human-like reasoning if trained properly [5]. They are capable of processing large-scale and complex data and establishing logical and syntactical relations among them [21]. Transformer models are also efficient in data analysis and draw logical conclusions from there [22]. Such models can handle both structured and unstructured data, combine multiple contexts, including both numerical and descriptive data, and make decisions [23]. These properties make transformer models an appropriate tool for enterprise system automation.

2.5. Research Gaps

The findings from the literature review reveal a set of research gaps. The primary objective of this study is to fill these gaps. The gaps identified in this study are listed as follows:
  • Lack of efficient adaptation of transformer models in enterprise systems to automate manual efforts.
  • Underexplored scope of cost minimization through process optimization in enterprise systems.
  • Absence of a feasibility analysis through real-world exploration of the transformer model as a substitute for manual efforts in enterprise systems.
The findings from the systematic literature review presented in this section play a foundational role in this study. The contributions presented in the introduction are justified through the relevant research gaps discovered in the literature review. The proposed EnterpriseAI framework aims to address these gaps by leveraging advanced NLP techniques and scalable design principles.

3. Methodology

3.1. Dataset Preparation and Preprocessing

Enterprise systems involve multiple different types of processes. The dynamic workflow associated with the systems usually produces a massive volume of structured and unstructured data [24]. However, these data vary from system to system. This study has been conducted on an enterprise system associated with a manufacturing industry located in different geographical locations that operate in harmony through the enterprise system. To enable the proposed EnterpriseAI framework, it is essential to develop a dataset that reflects the complex nature of the enterprise system’s operational process. A specialized dataset has been prepared and preprocessed to ensure this issue is addressed and maintains relevance and applicability.

3.1.1. Dataset Formation

The dataset was constructed from enterprise-specific scenarios, including workflow descriptions, resource allocation plans, and risk mitigation strategies. Data were collected from publicly available enterprise case studies, anonymized internal reports, and industry white papers. All of these documents are related to manufacturing enterprise systems. The collected content was categorized into three categories, which are presented in the list below.
  • Workflow data ( W ): detailed descriptions of enterprise workflows.
  • Resource data ( R ): information on resource allocation and usage.
  • Risk dat ( M ): case studies addressing risk factors and mitigation strategies.
These can be used to produce an initial informed dataset, and the categories have been merged together. The merging process follows the mathematical principle provided in Equation (1). Here, W p , R q , and M r denote individual entries in the workflow, resource, and risk datasets, respectively [25].
D = { W p , R q , M r p P , q Q , r R }
Training a transformer model through a transfer learning approach requires the dataset to be in a query–response format. The initial unified dataset prepared using Equation (1) does not have this structure. Later, in the preprocessing steps, the dataset was further transformed into query–response format through semi-manual intervention by following the concept presented in Equation (2) [26].
( u , v ) = { ( W p , A p ) , ( R q , S q ) , ( M r , T r ) p P , q Q , r R }
In Equation (2), A p , S q , and T r represent the annotations or responses associated with W p , R q , and M r , respectively. This structured dataset allows the transformer model to effectively learn enterprise-specific contextual relationships. Table 1 summarizes the key properties of the dataset, including the number of documents in each category, document types, structure, and word counts.
Table 1 shows that the dataset includes a total of 750 documents categorized into workflow, resource, and risk data. The documents vary in structure, ranging from structured reports to unstructured case studies, ensuring a diverse dataset suitable for training the EnterpriseAI framework. The total word count of the dataset is 904,000, providing a comprehensive knowledge base for the model.

3.1.2. Data Normalization and Cleaning

To standardize the data, all text entries were normalized to lowercase using Unicode transformation rules, ensuring uniformity without affecting semantic meaning [27]. The normalization process is expressed in Equation (3), where N ( T ) denotes the normalized text, and Δ is the transformation value applied to convert uppercase characters into lowercase.
N ( T ) = { τ k τ k = τ k , if τ k is lowercase ; τ k = τ k + Δ , if uppercase }
After normalization, irrelevant characters and noise were removed using Equation (4), where R represents the set of irrelevant elements excluded from the text.
C ( T ) = { τ k τ k T , τ k R }

3.1.3. Tokenization and Embedding Formation

To prepare the data for input into the transformer model, tokens were generated using a hybrid byte-pair encoding method. The tokenization process is expressed in Equation (5).
ζ ( T ) = k = 1 n { τ k 1 , τ k 2 , , τ k m W k T , m = len ( W k ) }
The dataset contains a total of 904,000 words spread across 1808 pages. With an average word length of 4.5 characters and an estimated token length of 3.0 characters, the dataset comprises approximately 1,356,000 tokens. On average, each page contributes around 750 tokens, ensuring an even distribution of data. The tokens were subsequently embedded into dense vector representations, defined in Equation (6).
v k = E ( τ k ) τ k ζ ( T )
In Equation (6), E ( τ k ) maps tokens τ k into a continuous vector space. This transformation facilitates efficient processing by the transformer model, enabling it to capture semantic and contextual relationships effectively.

3.1.4. Dataset Splitting

The dataset was divided into training, validation, and testing subsets to train the transformer model used in the EnterpriseAI framework and evaluate its performance. The splitting was performed following the widely used ratio of 70:15:15 for training, validation, and testing, respectively, as recommended in machine learning best practices [28]. During preliminary experiments, alternative splitting ratios, such as 60:20:20 and 80:10:10, were considered. However, the 70:15:15 ratio yielded the most stable performance in terms of convergence speed and evaluation metrics, making it the optimal choice for this study. The dataset comprises 750 documents, categorized into three main groups: workflow data ( W ), resource data ( R ), and risk data ( M ). These documents span a total of 1808 pages, containing 904,000 words and approximately 1,356,000 tokens. Table 2 summarizes the number of documents and tokens allocated to each subset.
This thoroughly prepared dataset ensures that the EnterpriseAI framework can learn and adapt to the complex requirements of enterprise systems.

3.2. Transformer Model Architecture

The core of the proposed EnterpriseAI framework is a custom transformer model designed to handle the complexities and interdependencies of enterprise systems. The model architecture has been optimized to ensure scalability, efficiency, and adaptability to diverse enterprise scenarios.

3.2.1. Model Design

The transformer model presented in this study has been developed through a series of comparative analyses. The finalized version summarized in Table 3 is the optimal architecture that produces the most reliable output with the least computational resources [29].

3.2.2. Input Embedding Layer

The dimension of the input embedding layer is 15,000 × 128. It has been designed to map the input tokens to the dense vectors. That means it converts tokenized sequences into dense vector representations. The embedding process is expressed in Equation (7) [30].
e z = E [ ψ z ] ψ z Ψ
Here, E is the embedding matrix, and ψ z represents the z-th token in the input sequence Ψ . The output of the embedding layer is a matrix X R L × 256 , where L is the sequence length, defined in Equation (8).
X = [ e 1 ; e 2 ; ; e L ]

3.2.3. Positional Encoding Layer

It is essential to preserve the position of the tokens to maintain the meaning of the features. This responsibility is taken care of by the positional encoding layer. To preserve token sequence order, a positional encoding vector is added to each token embedding. The positional encoding process is defined in Equation (9) [31], where p z [ k ] represents the k-th dimension of the positional vector for the z-th token, and d model is the model’s embedding dimension [31].
p z [ k ] = sin z 10 , 000 k / d model if k is even , cos z 10 , 000 k / d model if k is odd ,

3.2.4. Transformer Encoder Block

Each encoder block contains a multi-head attention mechanism, layer normalization, and a feed-forward network (FFN). The multi-head attention mechanism aggregates contextual information from the input sequence by computing attention scores. For each token ψ i , the query ( q i ), key ( k j ), and value ( v j ) vectors are computed as q i = W Q e i , k j = W K e j , v j = W V e j . The attention score α i j is then calculated using the scaled dot product expressed in Equation (10) [31].
α i j = softmax q i k j d model
The output from the attention mechanism is a weighted sum. The value of the weighted sum is calculated using Equation (11).
z i = j = 1 L α i j v j
The feed-forward neural network used in the transformer model applies two linear transformations with a ReLU activation in between which is governed by Equation (12), where W 1 , W 2 , b 1 , and b 2 are the trainable parameters of the FFN.
y i = ReLU ( W 1 z i + b 1 ) W 2 + b 2

3.2.5. Output Layer

The output layer is responsible for delivering the output generated by the transformer model. It is another embedding layer that maps the encoder’s output to logits, which are converted into probabilities using the Softmax function. The process follows the mathematical principle presented in Equation (13), where o i is the logit for token ψ i , and V is the vocabulary size.
P ( ψ i ) = exp ( o i ) j = 1 V exp ( o j )

3.3. Training the EnterpriseAI Transformer

The training process of the EnterpriseAI transformer model focused on fine-tuning a pre-trained transformer architecture to adapt it to enterprise-specific tasks. The training leveraged transfer learning to expedite the process and enhance the model’s performance on the prepared dataset. The learning curve illustrated in Figure 1 shows how effectively the transformer model learned from the dataset during the training process [32].

3.3.1. Fine-Tuning Objective

The fine-tuning objective was to minimize the discrepancy between the model’s predicted output and the true output from the dataset. This was achieved by optimizing the cross-entropy loss function, as defined in Equation (14) [33].
L = 1 N i = 1 N j = 1 V P ( ω j | ω 1 : i 1 ) log P ^ ( ω j | ω 1 : i 1 )
In Equation (14), P ( ω j | ω 1 : i 1 ) represents the true probability of token ω j , given the preceding tokens, and P ^ ( ω j | ω 1 : i 1 ) is the model’s predicted probability.

3.3.2. Optimization Strategy

The Adaptive Moment Estimation (ADAM) optimizer with weight decay was used to minimize the loss. The update rule for the model’s parameters Θ t at step t is given by Equation (15) [34].
Θ t + 1 = Θ t η · M t V t + ϵ
Here, η is the learning rate, M t and V t are the first and second moment estimates of the gradient, respectively, and ϵ is a small constant to prevent division by zero.

3.3.3. Learning Rate Scheduler

A learning rate scheduler was employed to adjust the learning rate dynamically during training. The scheduler followed a warm-up phase, gradually increasing the learning rate for the initial steps before decaying it. The learning rate at step t is defined in Equation (16) [35].
η t = η 0 · min 1 , t t warmup · 1 t
In Equation (16), η 0 is the initial learning rate, t warmup is the number of warm-up steps, and t is the current training step.

3.3.4. Evaluation During Training

During training, the model’s performance was evaluated on the validation set after every epoch. Metrics such as accuracy, precision, recall, and perplexity were computed to monitor progress. The perplexity metric, a measure of the model’s uncertainty in generating predictions, is defined in Equation (17) [36].
Perplexity = 2 H ( P , P ^ )
Here, H ( P , P ^ ) is the cross-entropy between the true probability distribution P and the predicted distribution P ^ .

4. Enterprise Use Case Integration

The EnterpriseAI framework is designed to seamlessly integrate the transformer model trained on enterprise-related data into existing enterprise systems and address real-world use cases. The literature review in Section 2 shows that the utilization of AI models in real-world applications for enterprise systems is a significant research gap. This section fills up this gap by seamlessly integrating the transformer model into enterprise systems.

4.1. Integration Architecture

The integration architecture has been developed using layered and modular approaches. A transformer model trained for one enterprise system may not be suitable for another, so it will need to be retrained using the relevant data. The integration architecture has been developed to accept any trained transformer model that starts working immediately. The integration architecture has been developed with three layers, as illustrated in Figure 2.
The first layer of the integration architecture is the data sources, which can accommodate multiple data sources simultaneously. This layer cleans the data, normalizes them, and covers them into tokens, making them suitable for the second layer. The second layer is the middleware layer, which maintains communication between the transformer and the framework. The third and final layer is the visualization layer, which consists of the dashboards, reporting tools, and decision-making workflows.

4.2. Key Enterprise Applications

The proposed EnterpriseAI has been applied in four different applications. The first application is financial forecasting and risk analysis, the second one is supply chain optimization, the third application is customer experience enhancement, and the fourth application is human resource management. Each application involves different operational variables that the model processes to generate cost-optimized decisions. For example, in supply chain optimization, the key variables are supplier reliability ( S r ), inventory levels ( I l ), logistics cost ( L c ), and demand variability ( D v ). The EnterpriseAI framework processes historical financial data and real-time market trends to generate accurate forecasts and identify potential risks. The transformer model captures intricate patterns in the data, providing actionable insights for financial planning and risk mitigation. EnterpriseAI enables real-time analysis of supply chain operations, identifying bottlenecks and optimizing logistics. The model analyzes shipment data, inventory levels, and supplier performance, suggesting cost-effective strategies to enhance operational efficiency. By analyzing customer feedback, support tickets, and interaction histories, the framework generates insights into customer sentiment and behavior. Enterprises use these insights to personalize customer experiences, improve satisfaction, and drive loyalty. The framework assists in workforce planning, recruitment, and retention strategies. It analyzes employee performance data, attrition rates, and industry benchmarks to recommend optimal HR strategies.

4.3. Implementation in Real-World Scenarios

The EnterpriseAI framework has been deployed in a pilot project within a manufacturing multinational corporation. The deployment focused on supply chain optimization and financial forecasting. Table 4 summarizes the outcomes achieved.

4.4. Input Variables and Relationship Identification

The effectiveness of the EnterpriseAI framework relies on identifying relationships between various enterprise data variables. The transformer model processes structured and unstructured data from multiple sources to derive meaningful cost-saving strategies. The following categories of input variables have been utilized in the cost optimization process:
  • Operational costs ( C o ): Employee workload, task complexity, and automation impact.
  • Resource utilization ( R u ): Labor allocation, equipment downtime, and material usage.
  • Supply chain factors ( S c ): Vendor performance, transportation logistics, and procurement efficiency.
  • Financial metrics( F m ): Revenue projections, risk exposure, and investment efficiency.
  • Customer experience ( CX e ): User satisfaction, service response time, and feedback trends.
These variables are analyzed through transformer attention layers, enabling EnterpriseAI to make cost-optimized recommendations dynamically.

4.5. Challenges and Benefits

The deployment of EnterpriseAI faced several challenges that were effectively addressed to ensure seamless integration and optimal performance. Data inconsistencies, such as missing or erroneous information within enterprise systems, were resolved using preprocessing and imputation techniques. System integration barriers with legacy systems were mitigated by developing custom APIs and middleware components, enabling smooth communication between EnterpriseAI and the existing infrastructure. Performance bottlenecks were addressed by optimizing inference pipelines through batching and asynchronous processing, ensuring the system could meet enterprise-scale demands. Despite these challenges, the deployment yielded significant benefits, including substantial cost savings in supply chain and financial operations, enhanced decision making through real-time analytics and accurate forecasts, improved customer satisfaction and employee retention rates, and remarkable scalability and adaptability across various departments and workflows.

5. Implementation

The proposed EnterpriseAI framework has been implemented on a robust computational infrastructure so that multiple experiments can be conducted. Moreover, the training requires high computational resources. The hardware used in this study and the software stack utilized have been illustrated in Figure 3.

5.1. Hardware Configuration and Software Stack

The hardware configuration for this project has been carefully selected to process a massive volume of data for the training and train the transformer model efficiently. Training a transformer model is time-consuming, and the model presented in this paper has been trained after multiple trials and error attempts. That is why hardware configuration with very high computational capability has been selected for this paper. The software stack has been carefully selected, which offers the flexibility to modify the transformer model as necessary. A list of the hardware configurations and software stacks is presented in Table 5.

5.2. Implementation Workflow

The implementation of EnterpriseAI followed a structured workflow to ensure efficient development and deployment. It has been presented in Figure 4. It starts with the data preprocessing. After that, the processed data are used to train the model. Before deployment, the inference pipeline has been developed. Finally, the model is deployed using Docker container on Kubernetes clusters.

5.3. Training and Inference Time

The training process for the EnterpriseAI model took approximately 29 h for 145 epochs on the specified hardware. The average inference time for a single query was 95 ms, demonstrating the framework’s suitability for real-time applications. The training and inference time along with other relevant data are presented in Table 6.

6. Performance Evaluation and Results

The performance analysis of the proposed EnterposeAI framework is categorized into two broad categories. The first category is related to the machine learning performance evaluation. At the heart of this framework is the transformer model, and the transformer model makes every intelligent decision the system makes. That is why the performance of it has been evaluated from the machine learning performance evaluation perspective. The second category is the performance evaluation, which involves the feasibility of the approach.

6.1. Evaluation Metrics

To evaluate the performance of the transformer model from the machine learning performance perspective, the accuracy, precision, recall, and F1-score were used. These performance metrics are defined in Equations (18), (19) and (20), respectively. These metrics are dependent on true positive (TP), true negative (TN), false positive (FP), and false negative (FN). These values have been obtained from the confusion metrics analysis [37,38].
Accuracy = T P + T N T P + T N + F P + F N
Precision = T P T P + F P
Recall = T P T P + F N
F 1 - Score = 2 · Precision · Recall Precision + Recall
Along with other evaluation metrics, the perplexity score has been used to evaluate the ability of the transformer model to generate meaningful results, which is defined in Equation (22), where H ( P , P ^ ) represents the cross-entropy between the true distribution P and the predicted distribution P ^ .
Perplexity = 2 H ( P , P ^ )

6.2. Confusion Matrix Analysis

There are a total of 271 pages in the test dataset, which consists of 135,600 words. To evaluate the performance of the proposed transformer model, 400 paragraphs were randomly picked from the testing dataset. After that, these paragraphs were manually labeled according to their subject into one of the following categories: Financial Forecasting, Risk Analysis, Supply Chain Optimization, Customer Experience, and Human Resource Management. After that, the ability of the transformer model to accurately classify the paragraphs was evaluated. The confusion matrix illustrated in Figure 5 shows the classification performance of the transformer model. According to the confusion matrix analysis, the overall accuracy, precision, recall, and F1-score are 91.09%, 91.10%, 91.09%, and 91.09%. These values indicate that the transformer model is well-trained and capable of understanding the content correctly.

6.3. k-Fold Cross-Validation

To further validate the performance of the transformer model, k f o l d cross-validation has been performed with k = 10 . The cross-validation has been performed on the test dataset. The findings have been presented in Table 7. The k f o l d cross-validation further validates the performance and consistency of the transformer model trained in this paper.
The consistency in performance discovered through the k-fold cross-validation has been visually presented in Figure 6 to explore the variations in performance at different folds. The value range in this figure is from around 90% to 91.8%. That means the variation is marginal, which again suggests that the proposed transformer model is capable of maintaining consistent performance.

6.4. Scalability and Resource Utilization

Depending on the demand and number of users, queries made by the enterprise systems can significantly vary. That is why the framework must be scalable. It is a mandatory requirement to consider the proposed framework as a feasible solution. The scalability is directly associated with the resource utilization. A scalability test has been performed that relates to resource utilization as well. The results obtained from the test have been presented in Table 8.
Figure 7 demonstrates the scalability range. The maximum response time of the system is 350 milliseconds, which is for the maximum number of queries. This is the threshold, and beyond it, the system fails to maintain acceptable performance. The system utilizes a moderate volume of computational resources to process 100 simultaneous queries. Although no significant performance deviation is noticeable for 101 to 200 queries, the GPU utilization touches 85% at this level. Going beyond it occupies 80% CPU and 90% of the GPU, leaving limited resources for the other core operations. That means the proposed system is scalable for up to 200 queries at a time.

6.5. Workflow Optimization

The EnterpriseAI framework significantly improved workflow efficiency across various enterprise functions. By automating repetitive tasks, streamlining resource allocation, and providing data-driven insights, the system optimized key performance indicators (KPIs) for workflow efficiency. Table 9 presents a comparison of workflow metrics before and after implementing the EnterpriseAI framework.
Figure 8 highlights substantial improvements in workflow metrics. Task completion time was reduced by 33.3%, while resource utilization efficiency increased by 30.8%. Furthermore, the error rate decreased significantly by 66.7%, contributing to a 50.0% improvement in project delivery rate. The customer satisfaction score also increased by 15.4%, demonstrating the positive impact of EnterpriseAI on overall workflow performance. By automating data processing and leveraging intelligent recommendations, EnterpriseAI allowed enterprise systems to operate more efficiently, reduce manual interventions, and achieve higher operational excellence.

6.6. Business Cost Optimization

The EnterpriseAI framework significantly contributes to optimizing business costs by automating manual tasks, minimizing resource wastage, and reducing error rates. By leveraging intelligent data processing and real-time insights, the system enables enterprises to cut operational expenses while improving overall efficiency. Table 10 highlights the cost optimization achieved in different operational areas.
Table 10 demonstrates the significant cost savings achieved across various operational areas. Manual effort costs were reduced by 58.3%, error and rework costs decreased by 66.7%, and inefficiencies in resource allocation saw a reduction of 60.0%. Additionally, the system shortened project durations by 30.0% and reduced unforeseen expenses related to risk mitigation by 50.0%.
These improvements highlight the system’s ability to optimize business costs effectively, providing enterprises with a scalable and cost-efficient solution for operational management.

7. Limitations and Future Directions

While the EnterpriseAI framework demonstrates significant potential in optimizing enterprise operations, several limitations and areas for improvement have been identified. This section outlines these limitations and discusses potential future directions to enhance the framework.

7.1. Limitations

7.1.1. Dataset Generalization

The current model relies on a dataset specifically tailored for enterprise use cases; while it is effective for the tested scenarios, the dataset lacks diversity across different industries and regions. This limitation may hinder the framework’s adaptability to niche or highly specialized enterprise domains.

7.1.2. Computational Complexity

The EnterpriseAI framework employs a transformer model with a large number of parameters, making it computationally intensive. The high resource requirements for training and inference may pose challenges for deployment in resource-constrained environments.

7.1.3. Interpretability

Like most transformer-based models, EnterpriseAI functions as a black-box system [39]. The lack of interpretability limits the ability to understand why specific decisions or predictions were made, which is critical for trust and transparency in enterprise applications.

7.1.4. Real-Time Adaptability

The framework is static in its current form, relying solely on pre-trained knowledge. It cannot dynamically adapt to real-time changes in enterprise operations or rapidly evolving market conditions, which could limit its effectiveness in highly dynamic environments.

7.1.5. Scalability Bottlenecks

While the framework performs well under moderate workloads, scalability to extremely high query loads introduces minor latency increases [40]. Further optimization is needed to ensure consistent performance in large-scale, high-demand enterprise systems.

7.2. Future Directions

7.2.1. Dataset Expansion and Diversity

Future work will focus on constructing a more diverse and comprehensive dataset that spans multiple industries, geographies, and operational contexts. This will improve the framework’s adaptability and generalization capabilities.

7.2.2. Model Compression and Optimization

To address computational complexity, techniques such as model pruning, quantization, and knowledge distillation will be explored. These techniques can reduce model size and resource requirements while maintaining performance.

7.2.3. Explainable AI (XAI) Integration

Integrating explainable AI techniques into the framework will enhance its interpretability. Approaches such as attention visualization, feature attribution, and decision tracing will be employed to provide insights into the model’s decision-making process.

7.2.4. Real-Time Learning and Adaptation

Incorporating online learning mechanisms will enable the framework to update its knowledge base in real-time. Techniques like incremental learning and continual learning will be explored to allow the model to adapt to dynamic enterprise environments.

7.2.5. Edge Deployment for Resource-Constrained Environments

Future iterations of the framework will explore deployment on edge devices, enabling it to operate efficiently in resource-constrained environments. Lightweight model architectures and optimized inference pipelines will be developed for this purpose.

7.2.6. Enhanced Scalability for High-Demand Applications

Scalability improvements will focus on advanced distributed computing techniques, including sharding, micro-batching, and asynchronous query processing. These enhancements will ensure consistent performance under high-demand scenarios.

7.2.7. Enterprise-Specific Customizations

Future versions of the framework will include customizable modules tailored to specific enterprise needs. This could involve fine-tuning the model for industry-specific terminology, workflows, and regulations.

7.3. Collaborative Research Opportunities

Collaboration with industry partners and academic institutions will be pursued to benchmark the framework against real-world datasets and scenarios. Such partnerships will also provide valuable feedback to refine and extend the system’s capabilities.

7.4. Long-Term Vision

The long-term vision for EnterpriseAI is to create a fully autonomous enterprise assistant that can seamlessly integrate with existing systems, adapt to real-time changes, and provide actionable insights with minimal human intervention. Achieving this vision will involve a combination of AI advancements, robust data pipelines, and close collaboration with stakeholders.
The identified limitations and proposed future directions provide a roadmap for advancing the EnterpriseAI framework, ensuring its continued relevance and effectiveness in addressing the evolving challenges of enterprise operations.

8. Conclusions

The EnterpriseAI framework presented in this paper offers a scalable, efficient, and intelligent solution for enterprise systems, addressing critical challenges in data-driven decision making, resource optimization, and operational efficiency. Built upon a transformer-based architecture, the framework demonstrates exceptional capabilities in handling complex enterprise processes, delivering real-time insights, and significantly reducing operational costs.
The performance evaluation validates the framework’s robustness and reliability, with high accuracy, precision, recall, and F1-scores across various tasks, including phase classification and priority prediction. The successful deployment in enterprise use cases, such as financial forecasting, supply chain optimization, and customer experience enhancement, highlights its practical relevance and business impact. Moreover, the framework’s scalability and efficient resource utilization make it well suited for diverse enterprise environments.
Despite its promising performance, the study acknowledges several limitations, including dataset generalization, computational complexity, and interpretability challenges. These limitations present opportunities for future research and development, focusing on enhancing adaptability, reducing resource requirements, and integrating Explainable AI techniques. Additionally, the proposed directions for real-time learning, edge deployment, and industry-specific customizations will further refine the framework and extend its applicability.
In conclusion, the EnterpriseAI framework represents a significant advancement in the application of AI to enterprise systems. By leveraging state-of-the-art transformer models, it provides a powerful tool for enterprises to optimize operations, improve decision making, and achieve strategic objectives. With continued research and development, this framework has the potential to redefine enterprise processes and contribute to the future of intelligent and autonomous enterprise systems.

Funding

This research received no external funding.

Data Availability Statement

The data supporting the reported results is proprietary and cannot be made publicly available. However, it can be accessed on reasonable request, subject to the agreement of a non-disclosure and non-commercial application policy.

Conflicts of Interest

Shinoy Vengaramkode Bhaskaran is employed by Zoom Video Communications.

Abbreviations

The following abbreviations are used in this manuscript:
AIartificial intelligence
SDLCsystem development life cycle
LLMlarge language model
NLPnatural language processing
GPUgraphics processing unit
CPUcentral processing unit
APIapplication programming interface
MLmachine learning
MDPIMultidisciplinary Digital Publishing Institute

References

  1. Solano, M.C.; Cruz, J.C. Integrating analytics in enterprise systems: A systematic literature review of impacts and innovations. Adm. Sci. 2024, 14, 138. [Google Scholar] [CrossRef]
  2. Fuad, K.; Li, P.; Maruping, L.; Mathiassen, L. An absorptive capacity framework for investigating enterprise system ecosystems: The role of connectivity and intelligence. Enterp. Inf. Syst. 2024, 18, 2330084. [Google Scholar] [CrossRef]
  3. Panigrahi, R.; Bele, N.; Panigrahi, P.K.; Gupta, B.B. Features level sentiment mining in enterprise systems from informal text corpus using machine learning techniques. Enterp. Inf. Syst. 2024, 18, 2328186. [Google Scholar] [CrossRef]
  4. Himeur, Y.; Elnour, M.; Fadli, F.; Meskin, N.; Petri, I.; Rezgui, Y.; Bensaali, F.; Amira, A. AI-big data analytics for building automation and management systems: A survey, actual challenges and future perspectives. Artif. Intell. Rev. 2023, 56, 4929–5021. [Google Scholar] [CrossRef]
  5. Raiaan, M.A.K.; Mukta, M.S.H.; Fatema, K.; Fahad, N.M.; Sakib, S.; Mim, M.M.J.; Ahmad, J.; Ali, M.E.; Azam, S. A review on large Language Models: Architectures, applications, taxonomies, open issues and challenges. IEEE Access 2024, 12, 26839–26874. [Google Scholar] [CrossRef]
  6. Lehmann, J.; Bhandiwad, D.; Gattogi, P.; Vahdati, S. Beyond boundaries: A human-like approach for question answering over structured and unstructured information sources. Trans. Assoc. Comput. Linguist. 2024, 12, 786–802. [Google Scholar] [CrossRef]
  7. Tomaszewski, R. A study of citations to STEM databases: ACM Digital Library, Engineering Village, IEEE Xplore, and MathSciNet. Scientometrics 2021, 126, 1797–1811. [Google Scholar] [CrossRef]
  8. Li, Z.; Rainer, A. Reproducible Searches in Systematic Reviews: An Evaluation and Guidelines. IEEE Access 2023, 11, 84048–84060. [Google Scholar] [CrossRef]
  9. Jiang, Y.; Jeusfeld, M.A.; Mosaad, M.; Oo, N. Enterprise architecture modeling for cybersecurity analysis in critical infrastructures-A systematic literature review. Int. J. Crit. Infrastruct. Prot. 2024, 46, 100700. [Google Scholar] [CrossRef]
  10. Shi, J. Adaptive change: Emerging economy enterprises respond to the international business environment challenge. Technovation 2024, 133, 102998. [Google Scholar] [CrossRef]
  11. Al-Assaf, K.; Alzahmi, W.; Alshaikh, R.; Bahroun, Z.; Ahmed, V. The relative importance of key factors for integrating Enterprise Resource Planning (ERP) systems and performance management practices in the UAE Healthcare Sector. Big Data Cogn. Comput. 2024, 8, 122. [Google Scholar] [CrossRef]
  12. Kulkarni, V.; Reddy, S.; Clark, T.; Proper, H. The AI-Enabled Enterprise. In The AI-Enabled Enterprise; Springer: Berlin/Heidelberg, Germany, 2023; pp. 1–12. [Google Scholar]
  13. Forcina, A.; Silvestri, L.; De Felice, F.; Falcone, D. Exploring Industry 4.0 technologies to improve manufacturing enterprise safety management: A TOPSIS-based decision support system and real case study. Saf. Sci. 2024, 169, 106351. [Google Scholar] [CrossRef]
  14. Nandanwar, H.; Katarya, R. Deep learning enabled intrusion detection system for Industrial IOT environment. Expert Syst. Appl. 2024, 249, 123808. [Google Scholar] [CrossRef]
  15. Garg, H.; Khan, M.I.; Yanhong, L.; Ibrar, M.; Nazif, F.; Latif, A. Selection of best enterprise resource planning system by using Hamy mean operator with complex spherical fuzzy information. Alex. Eng. J. 2024, 86, 494–512. [Google Scholar] [CrossRef]
  16. Rehman, Z.; Tariq, N.; Moqurrab, S.A.; Yoo, J.; Srivastava, G. Machine learning and internet of things applications in enterprise architectures: Solutions, challenges, and open issues. Expert Syst. 2024, 41, e13467. [Google Scholar] [CrossRef]
  17. Jawad, Z.N.; Balázs, V. Machine learning-driven optimization of enterprise resource planning (ERP) systems: A comprehensive review. Beni-Suef Univ. J. Basic Appl. Sci. 2024, 13, 4. [Google Scholar] [CrossRef]
  18. Parycek, P.; Schmid, V.; Novak, A.S. Artificial Intelligence (AI) and automation in administrative procedures: Potentials, limitations, and framework conditions. J. Knowl. Econ. 2024, 15, 8390–8415. [Google Scholar] [CrossRef]
  19. Jia, Y.; Wang, Z. Application of artificial intelligence based on the fuzzy control algorithm in enterprise innovation. Heliyon 2024, 10, e28116. [Google Scholar] [CrossRef]
  20. Kreutz, H.; Jahankhani, H. Impact of Artificial Intelligence on Enterprise Information Security Management in the Context of ISO 27001 and 27002: A Tertiary Systematic Review and Comparative Analysis. In Cybersecurity and Artificial Intelligence: Transformational Strategies and Disruptive Innovation; Springer: Cham, Switzerland, 2024; pp. 1–34. [Google Scholar]
  21. Zhang, H.; Shafiq, M.O. Survey of transformers and towards ensemble learning using transformers for natural language processing. J. Big Data 2024, 11, 25. [Google Scholar] [CrossRef]
  22. Onan, A.; Alhumyani, H.A. FuzzyTP-BERT: Enhancing extractive text summarization with fuzzy topic modeling and transformer networks. J. King Saud-Univ.-Comput. Inf. Sci. 2024, 36, 102080. [Google Scholar] [CrossRef]
  23. Li, B.; Jiang, G.; Li, N.; Song, C. Research on large-scale structured and unstructured data processing based on large language model. In Proceedings of the International Conference on Machine Learning, Pattern Recognition and Automation Engineering, Singapore, 7–9 August 2024; pp. 111–116. [Google Scholar]
  24. Zur Muehlen, M. Workflow-Based Process Controlling: Foundation, Design, and Application of Workflow-Driven Process Information Systems; Michael zur Muehlen: Berlin, Germany, 2004; Volume 6. [Google Scholar]
  25. Faruqui, N.; Thatoi, P.; Choudhary, R.; Roncevic, I.; Alqahtani, H.; Sarker, I.H.; Khanam, S. AI-Analyst: An AI-Assisted SDLC Analysis Framework for Business Cost Optimization. IEEE Access 2024, 12, 195188–195203. [Google Scholar] [CrossRef]
  26. Zhao, T.; Faruqui, N. EconoFormer: A Novel Macroeconomic Policy Analysis and Implementation Planner using Generative Transformer Model. IEEE Access 2024, 12, 184714–184725. [Google Scholar] [CrossRef]
  27. Sathupadi, K.; Achar, S.; Bhaskaran, S.V.; Faruqui, N.; Abdullah-Al-Wadud, M.; Uddin, J. Edge-cloud synergy for AI-enhanced sensor network data: A real-time predictive maintenance framework. Sensors 2024, 24, 7918. [Google Scholar] [CrossRef] [PubMed]
  28. Hossain, M.E.; Faruqui, N.; Mahmud, I.; Jan, T.; Whaiduzzaman, M.; Barros, A. DPMS: Data-driven promotional management system of universities using deep learning on social media. Appl. Sci. 2023, 13, 12300. [Google Scholar] [CrossRef]
  29. Acciaio, B.; Kratsios, A.; Pammer, G. Designing universal causal deep learning models: The geometric (hyper) transformer. Math. Financ. 2024, 34, 671–735. [Google Scholar] [CrossRef]
  30. Jung, M.; Lee, J.; Kim, J. A lightweight CNN-transformer model for learning traveling salesman problems. Appl. Intell. 2024, 54, 7982–7993. [Google Scholar] [CrossRef]
  31. Kazemnejad, A.; Padhi, I.; Natesan Ramamurthy, K.; Das, P.; Reddy, S. The impact of positional encoding on length generalization in transformers. Adv. Neural Inf. Process. Syst. 2024, 36, 24892–24928. [Google Scholar]
  32. Viering, T.; Loog, M. The shape of learning curves: A review. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 7799–7819. [Google Scholar] [CrossRef]
  33. Ho, Y.; Wookey, S. The real-world-weight cross-entropy loss function: Modeling the costs of mislabeling. IEEE Access 2019, 8, 4806–4813. [Google Scholar] [CrossRef]
  34. Zhang, W.; Niu, L.; Zhang, D.; Wang, G.; Farrukh, F.U.D.; Zhang, C. Hw-adam: Fpga-based accelerator for adaptive moment estimation. Electronics 2023, 12, 263. [Google Scholar] [CrossRef]
  35. Wen, L.; Li, X.; Gao, L. A new reinforcement learning based learning rate scheduler for convolutional neural network in fault classification. IEEE Trans. Ind. Electron. 2020, 68, 12890–12900. [Google Scholar] [CrossRef]
  36. Colla, D.; Delsanto, M.; Agosto, M.; Vitiello, B.; Radicioni, D.P. Semantic coherence markers: The contribution of perplexity metrics. Artif. Intell. Med. 2022, 134, 102393. [Google Scholar] [CrossRef]
  37. Zhou, J.; Gandomi, A.H.; Chen, F.; Holzinger, A. Evaluating the quality of machine learning explanations: A survey on methods and metrics. Electronics 2021, 10, 593. [Google Scholar] [CrossRef]
  38. Hicks, S.A.; Strümke, I.; Thambawita, V.; Hammou, M.; Riegler, M.A.; Halvorsen, P.; Parasa, S. On evaluation metrics for medical applications of artificial intelligence. Sci. Rep. 2022, 12, 5979. [Google Scholar] [CrossRef] [PubMed]
  39. Shi, Y.; Han, Y.; Tan, Y.a.; Kuang, X. Decision-based black-box attack against vision transformers via patch-wise adversarial removal. Adv. Neural Inf. Process. Syst. 2022, 35, 12921–12933. [Google Scholar]
  40. Li, S.; Jin, X.; Xuan, Y.; Zhou, X.; Chen, W.; Wang, Y.X.; Yan, X. Enhancing the locality and breaking the memory bottleneck of transformer on time series forecasting. Adv. Neural Inf. Process. Syst. 2019, 32, 5243–5253. [Google Scholar]
Figure 1. The learning curve of the proposed transformer model.
Figure 1. The learning curve of the proposed transformer model.
Computers 14 00106 g001
Figure 2. A simplified version of the integration architecture. The blue and green lines represent training and validation accuracy, respectively, while the red and purple lines indicate training and validation loss.
Figure 2. A simplified version of the integration architecture. The blue and green lines represent training and validation accuracy, respectively, while the red and purple lines indicate training and validation loss.
Computers 14 00106 g002
Figure 3. The overview of the hardware and software stack used in this experiment.
Figure 3. The overview of the hardware and software stack used in this experiment.
Computers 14 00106 g003
Figure 4. The implementation workflow from data preprocessing to deployment.
Figure 4. The implementation workflow from data preprocessing to deployment.
Computers 14 00106 g004
Figure 5. Confusion matrix to evaluate the classification performance of the transformer model.
Figure 5. Confusion matrix to evaluate the classification performance of the transformer model.
Computers 14 00106 g005
Figure 6. K-fold cross-validation performance visualization.
Figure 6. K-fold cross-validation performance visualization.
Computers 14 00106 g006
Figure 7. Scalability analysis with respect to the resource utilization.
Figure 7. Scalability analysis with respect to the resource utilization.
Computers 14 00106 g007
Figure 8. The workflow optimization comparison before and after using the EnterpriseAI. The underlying data supporting this figure is publicly available at https://drive.google.com/drive/folders/1PQmgN5eMX1GG8PcT-vu5oumjc3RMHAHh?usp=sharing (accessed on 2 January 2025).
Figure 8. The workflow optimization comparison before and after using the EnterpriseAI. The underlying data supporting this figure is publicly available at https://drive.google.com/drive/folders/1PQmgN5eMX1GG8PcT-vu5oumjc3RMHAHh?usp=sharing (accessed on 2 January 2025).
Computers 14 00106 g008
Table 1. Properties of the EnterpriseAI dataset.
Table 1. Properties of the EnterpriseAI dataset.
CategoryNo. of DocumentsDocument TypeStructureAvg. Words/DocTotal Words
Workflow Data ( W )320Reports, GuidelinesStructured1200384,000
Resource Data ( R )250Allocation Plans, LogsSemi-structured1000250,000
Risk Data ( M )180Case Studies, ReportsUnstructured1500270,000
Total750MixedMixed-904,000
Table 2. Dataset splitting summary.
Table 2. Dataset splitting summary.
SubsetDocumentsPagesWordsTokens
Training ( 70 % )5251266632,800949,200
Validation ( 15 % )112271135,600202,200
Testing ( 15 % )113271135,600202,200
Total7501808904,0001,356,000
Table 3. Optimized transformer model architecture for EnterpriseAI.
Table 3. Optimized transformer model architecture for EnterpriseAI.
LayerTypeHeadsDimensionsDescription
1Input Embedding-15,000 × 128Maps input tokens to dense vectors
2Positional Encoding-128 × 128Adds positional information to embeddings
3–10Transformer Encoder Block--Comprises multi-head attention and FFN layers
Details of Transformer Encoder Block
3a–10aMulti-Head Attention4 heads128 × 128Computes attention scores across input sequences
3b–10bLayer Normalization-128Normalizes outputs of the attention mechanism
3c–10cFeed-Forward Network-128 → 512 → 128Applies point-wise transformations
3d–10dDropout--Prevents overfitting during training
11Output Embedding-128 × 15,000Maps encoder output to vocabulary logits
12Softmax Layer-15,000Converts logits into probabilities
Table 4. Business cost optimization metrics.
Table 4. Business cost optimization metrics.
Cost ComponentBefore (USD)After (USD)Savings (%)Key Input Variables Used
Manual Effort60,00025,00058.3Task duration, workload,
automation impact
Error and Rework
Costs
30,00010,00066.7Error rate, correction costs,
automation reliability
Resource Allocation
Inefficiency
20,000800060.0Workforce utilization,
demand forecasting
Extended Project
Duration
50,00035,00030.0Task dependency,
timeline prediction
Risk Mitigation40,00020,00050.0Risk factors, incident frequency,
mitigation success
Supply Chain
Optimization
45,00022,00051.1Supplier reliability, inventory levels,
logistics cost
Total Costs245,000120,00051.0-
Table 5. Hardware and software configurations for EnterpriseAI implementation.
Table 5. Hardware and software configurations for EnterpriseAI implementation.
CategorySpecifications
Hardware Configuration
ProcessorIntel Xeon Gold 6258R, 28 cores at 2.70 GHz
GPUNVIDIA A100 Tensor Core with 40 GB memory
Memory768 GB DDR4 RAM
Storage4 TB NVMe SSD + 12 TB HDD
Network10 Gbps Ethernet connection
Software Stack
Programming LanguagePython 3.9
Frameworks and LibrariesPyTorch 1.12, Hugging Face Transformers,
NumPy, Pandas 1.5, Scikit-learn 1.2.2
Development ToolsJupyter Notebook 6.5.4, Docker 24.0.2
Orchestration PlatformKubernetes 1.26.3
Monitoring ToolsPrometheus 2.44, Grafana 9.5.2
Table 6. Training and inference time analysis for EnterpriseAI framework.
Table 6. Training and inference time analysis for EnterpriseAI framework.
MetricTime/Performance
Total Training Time (145 epochs)29 h
Average Training Time per Epoch12 min
Average Inference Time per Query95 ms
Peak Hardware Utilization (GPU)78%
Peak Hardware Utilization (CPU)42%
Memory Usage During Inference2.5 GB
Maximum Query Load During Testing120 queries/second
Table 7. K-fold cross-validation results.
Table 7. K-fold cross-validation results.
FoldAccuracy (%)Precision (%)Recall (%)F1-Score (%)
190.890.990.790.8
291.291.391.291.2
391.091.191.091.0
491.491.491.391.3
591.191.091.191.1
691.391.291.291.2
790.991.090.991.0
891.291.391.291.2
991.091.191.091.0
1091.591.591.491.5
Average91.191.291.191.1
Table 8. Scalability analysis of EnterpriseAI framework.
Table 8. Scalability analysis of EnterpriseAI framework.
Query LoadResponse Time (ms)CPU Utilization (%)GPU Utilization (%)
Low (1–10)852050
Medium (11–50)1003565
High (51–100)1305075
Very High (101–200)1806585
Extreme (201–500)2508090
Maximum Stress (>500)3509595
Table 9. Workflow optimization metrics before and after EnterpriseAI implementation.
Table 9. Workflow optimization metrics before and after EnterpriseAI implementation.
MetricBeforeAfterImprovement (%)
Task Completion Time (hours)12833.3
Resource Utilization Efficiency (%)658530.8
Error Rate (%)15566.7
Project Delivery Rate (per month)101550.0
Customer Satisfaction Score789015.4
The underlying data supporting this table are publicly available at https://drive.google.com/drive/folders/1PQmgN5eMX1GG8PcT-vu5oumjc3RMHAHh?usp=sharing (accessed on 2 January 2025).
Table 10. Business cost optimization metrics.
Table 10. Business cost optimization metrics.
Cost ComponentBefore EnterpriseAI (USD)After EnterpriseAI (USD)Cost Savings (%)
Manual Effort60,00025,00058.3
Error and Rework Costs30,00010,00066.7
Resource Allocation Inefficiency20,000800060.0
Extended Project Duration50,00035,00030.0
Risk Mitigation (Unforeseen Expenses)40,00020,00050.0
Total Costs200,00098,00051.0
The underlying data supporting this table are publicly available at https://drive.google.com/drive/folders/1PQmgN5eMX1GG8PcT-vu5oumjc3RMHAHh?usp=sharing (accessed on 2 January 2025).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Bhaskaran, S.V. EnterpriseAI: A Transformer-Based Framework for Cost Optimization and Process Enhancement in Enterprise Systems. Computers 2025, 14, 106. https://doi.org/10.3390/computers14030106

AMA Style

Bhaskaran SV. EnterpriseAI: A Transformer-Based Framework for Cost Optimization and Process Enhancement in Enterprise Systems. Computers. 2025; 14(3):106. https://doi.org/10.3390/computers14030106

Chicago/Turabian Style

Bhaskaran, Shinoy Vengaramkode. 2025. "EnterpriseAI: A Transformer-Based Framework for Cost Optimization and Process Enhancement in Enterprise Systems" Computers 14, no. 3: 106. https://doi.org/10.3390/computers14030106

APA Style

Bhaskaran, S. V. (2025). EnterpriseAI: A Transformer-Based Framework for Cost Optimization and Process Enhancement in Enterprise Systems. Computers, 14(3), 106. https://doi.org/10.3390/computers14030106

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop