Tech Trend Analysis System: Using Large Language Models and Finite State Chain Machines
Abstract
:1. Introduction
- Section 2 introduces terminology for Large Language Models are described as neural networks trained on vast datasets to process text data and learn language patterns. Some notions revisited in this section include self-attention and transformer blocks, limitations induced by various parameters, techniques such as prompt engineering, Retrieval-Augmented Generation (RAG), and Large Language Model Agents.
- In Section 3 we define Finite State Chain Machine, a new computational model designed to capture state-based, directed calls of LLMs based on sequential inputs. FSCMs with polynomially bounded transition functions prove to be non-universal, but become so when arbitrary Turing computable functions are used.
- The Large-Scale Text Analytics System is presented in Section 4. An intelligent decision-support architecture for extracting insights from unstructured data is presented. It addresses the limitations of LLMs with static knowledge bases by proposing integration with an external Long-Term Memory (LTM), specifically a knowledge graph, for dynamic information storage and retrieval. The system architecture includes components like the Transcript Receiver, FSCM with LLM support, Graph Database, Knowledge Retrieval Module, Clustering Module, and Data Visualization Module. Heavy-duty tasks for the LLM involve translation, summarization, entity recognition, relation extraction, and entity disambiguation. The process includes structuring data into JSON, error correction, and consolidating entities using similarity metrics. Data are stored in a graph database, which allows semantic searches.
- Section 5, A Real Application, Finding Technological Trends, describes the application of the system to identify R&D trends by using earnings call transcripts of top tech companies. The process involves extracting data into JSON files using LLMs, which are then loaded into a knowledge graph. An algorithm for trend discovery is presented and several examples of identified trends and types are provided.
- Section 6 provides an overview of the research and the steps of the proposed trend analysis process, from data collection to trend identification using an LLM. It states the effectiveness and scalability of the developed method. Future lines of research, such as multimodal embeddings, are suggested.
2. Large Language Models
3. Finite State Chain Machine
- is a finite set of states;
- is the alphabet of M;
- is the initial state;
- is the initial history string;
- is a finite set of contexts, , , corresponding to the states of Q;
- is a finite set of computable functions corresponding to the states of Q, , , which maps a tuple containing a context string, an input string, and a history string to a pair consisting of an output string and a next state.
- M starts in the initial state . In this case, the context string is , the initial history string is , and the first input string is .
- At each step, assuming that q is the current state of M, c is the context string associated with the current state, w is the current input string to be processed from the list , and h is the history string, then M performs as follows:
- •
- M computes the new output string and the next state using the function f of the current state: . If f is not defined for , then M halts, rejecting the list of input strings.
- •
- M updates the history string h with the newly computed output, that is, .
- •
- M changes its state from q to p.
- Step 2 is repeated until all input strings have been processed (a successful computation) or it halts by rejecting them (as described above). In a successful computation, M returns the last computed output string.
- ;
- ;
- where the functions are defined by the following:
- The list of input strings ;
- Initial history string .
- 1.
- M starts in state , the context being .
- 2.
- M processes the first input string :
- M computes
- M updates history string .
- M performs a transition from state to state .
- 3.
- M processes in state , the context being now :
- M computes
- M update history .
- M performs the transition back to state .
- 4.
- Computation stops, all the input strings being processed.
- concatenation function, which concatenates two strings and ;
- reversal function, which reverses a given string;
- palindrome check function, which checks if a string s is a palindrome;
- word count function, which counts the number of words in a string;
- string compression function, which compresses a string s by replacing consecutive repeated characters with the character followed by the count.
- However, it is worth mentioning that if one considers arbitrary Turing computable functions for then the model becomes computationally universal.
4. Large-Scale Text Analytics System
- Transcript Receiver: Responsible for collecting and preprocessing Earning call transcripts or other texts from various sources.
- Finite State Chain Machine with LLM support: A modular, state-based processing framework that performs tasks, such as Translation, Summarization, Entity recognition, Entity filtering, Relation extraction, and Structuring data.
- Graph Database: Stores the extracted information in a graph structure to enable efficient querying and analysis.
- Knowledge Retrieval Module: Provides an interface for querying and retrieving specific information from the graph database.
- Clustering Module: Identifies related concepts across several texts.
- Data Visualization Module: Offers visualization tools to present insights and trends in an intuitive and user-friendly manner.
- Translating and summarizing the input text that contains relevant information on any given topic. Translation is required as input texts regarding the same subject might come from different sources, so one has to standardize the content. Summarization is meant to remove unnecessary details and obtain brief content.
- Finding entities and relations among them and representing them in a structured format.
- Entity disambiguation. In this case, the LLM is asked to identify multiple entities that actually refer to the same concept or subject and merge them into a single, consistent representation.
- chunk translation into English and summarization (the objective is to uniformize the input text and obtain a brief version of it). Correspondingly, the LLM model is fed with a prompt composed of the context c1 concatenated with the chunk:c1 = “Translate the text in English and build a very brief and concise bullet list version by including only the essential and relevant information. You must use the complete names for the identified entities such that each statement in the list could be self understandable and will refer to the same entities as the previous ones. The output must have around 150 words (plus or minus 10 words). Use simple telegraphic and logically connected sentences in plain English. Obey strictly this requirements. Here is the text:”
- For a summarized English version of a chunk, we produce a structured version of it by identifying interacting entities and relations among them. To this aim, the LLM model is fed with a prompt composed of the context c2 concatenated with the summarized text:c2 = “Analyze the given text and break it down into individual statements to build a JSON file with a ‘brief’ key summarizing the text, and an ‘assertions’ key. The assertions is a list where each element has the key ‘stm’ stating the fact or opinion, ‘id’ with a unique ID (’"+id+"_’ + counter), ‘summary’ with a brief summary, ‘domain’ with a listing of relevant comma-separated domains, ‘keywords’ with four relevant comma-separated keywords, and ‘knowledge’ in a structured format with all the entities (defined by ‘name’, ‘type’, and ‘properties’) and at least two relationships among the found entities (defined by ‘source’:‘type’ entity, ‘relation’, and ‘destination’:‘type’ entity) describing the reason why the corresponding entities are linked. Once entities are detected, they must be disambiguated to distinguish between entities with similar or identical names. Identify potential ambiguities among these entities, considering contextual clues such as their relationships with other nodes. Use your understanding of domain-specific knowledge to disambiguate any ambiguous entities. In relationships you have to only use the discovered entities and nothing else. Produce only the JSON output following this structure strictly. The text is:”Here, the ID represents a unique identifier which is composed by an id of the analyzed text and a counter.
{"stm": "The company saw strong performance across its portfolio and is excited about early traction in generative AI.", "id": "a32_3", "summary": "Company’s performance and interest in AI", "domain": "Business, Technology", "keywords": "Portfolio Performance, Generative AI, Traction, Excitement", "knowledge": { "entities": [{"name": "The Company", "type": "organization"}, {"name": "Generative AI", "type": "concept", "properties": {"field": "AI"}}], "relationships": [{"source": {"name": "The Company", "type": "organization"}, "relation": "sees strong performance in", "destination": {"name": "Portfolio", "type": "concept"}}, {"source": {"name": "The Company", "type": "organization"}, "relation": "is excited about", "destination": {"name": "Generative AI", "type": "concept"}}] }}
{“name”: “ABC Corp”, “type”: “company”} {“name”: “ABC Co.”, “type”: “organization”}
- then, the outcome of applying the described method is only one entity:
{“name”: “ABC Corp”, “type”: “company”}
- the keywords: several words which are relevant for the relation;
- the domain: the field/scope of the relation;
- the statement: a short sentence that briefly describes the relation;
- the source: a string representing the source of information (e.g., the filename containing the analyzed information);
- the timestamp: the moment when this relation was established;
- the index: the order of the relation in the input JSON file(s).
- [Entity ] [Relation ] [Entity ]. The assertion is related to domain [Domain ] and some relevant keywords are [Keyword ]. The statement describing the relation is: [Statement ].
- …
- [Entity ] [Relation ] [Entity ]. The assertion is related to domain [Domain ] and some relevant keywords are [Keyword ]. The statement describing the relation is: [Statement ].
- Generate a brief narrative that summarizes the key findings incorporating relevant details about [input word]. Ensure that the story flows logically and provides context for future exploration.
- Using SBERT (see [38]) one can generate dense vector representations for all these texts/sentences which further allows to perform efficient similarity comparison between them. In this way, one may automatically discover related concepts, similar ideas, or patterns and trends.
- Using the LLM and a proper context. For example, the input prompt might be the following:Given the texts [text/sentence] and [text/sentence] provide a list of common related concepts or ideas if any.
stream = ollama.generate( #model=‘deepseek-r1:70b’, model=‘llama3.3’, prompt=mycontext +"\""+mytext_english+"\"", stream=False, options={‘temperature’: 0, ‘num_ctx’: 4096} )
5. A Real Application, Finding Technological Trends
- Step 1. Finding the entity node(s) in the knowledge graph that exhibit the highest in or out-degree.
- (a)
- Use K-Means Clustering to group the anchor entities based on their attribute values (say the label attribute) into distinct clusters by using adaptive silhouette analysis (which is used to determine the optimal number of clusters).For example, assuming that the number of anchor entities is 200 (where the number of analyzed earning calls transcripts is 35), the algorithm returned the following set of labels:By using the silhouette score one can find that the optimal number of clusters is 3:
- (b)
- Select the representative anchors by sorting the entities within each cluster based on their combined in–out degree and using a cyclic top-k selector (that is, we are cycling through all clusters, selecting at each passing the highest value element from each cluster until we obtain a list of k entities).For example, we used , a number large enough to ensure that each company will be represented by at least one anchor entity.
- Step 2. Contextually relevant subgraph extraction.
- Step 3. Assembling the core document.
- Step 4. Identification of latent topics in the core document.
{"trends": [{"trend_name": "Artificial Intelligence", "type": "Technological"}, {"trend_name": "AI Adoption", "type": "Technological"}, {"trend_name": "Cloud Growth", "type": "Digital"}, {"trend_name": "Cybersecurity Consolidation", "type": "Security"}]}
- Cluster 1: [Autonomy Advancement, AI Opportunity, Artificial Intelligence, AI Innovation, AI Adoption, AI Leadership, AI Growth, AI Search, AI Models, Agentic AI, Small AI, Custom AI, Self Driving Cars, AI Training, AI Infrastructure, AI Platform, AI Partnerships, AI Agent Collaboration, AI Services, …]
- Cluster 2: [Power Electronics, Silicon Photonics, Advanced IoT, Custom Hardware, Edge Computing, High-Speed Networking, Hybrid Integration, Hardware Advancements, High Performance Computing, Advanced Process, Quantum Computing, Programmable Networks, Silicon Software Integration, 3D Dram, Advanced Chip, …]
- Cluster 3: [Cloud Infrastructure, Cloud Adoption, Cloud Expansion, Multicloud Expansion, Cloud Growth, Cloud Computing, Cloud Migration, Cloud Optimization, Cloud Transformation, Cloud Computing Decline, Multicloud Adoption, Cloud Productivity, Cloud Performance, Cloud Integration, Cloud Automation, Crosscloud Solutions, …]
- Cluster 4: [IoT Advancements, 5G Adoption, Data Innovation, Platformization, Robotics Growth, Data Analytics, Innovative Technologies, Cybersecurity Advancement, Cybersecurity Demand, Cybersecurity Growth, Wireless Access, 5G Development, 5g Content, Digital Transformation, Smartphone Market, IoT Growth, Robotics Expansion, Robotics Adoption, …]
- Cluster 5: [Cloud Growth, Cloud Computing, Cloud Expansion, Cloud Transformation, Cloud Services, Cloud Transition, Cloud Native, Cloud Adoption, Cloud Infrastructure, Cloud Security, Cloud Gaming, Cloud Automation, Cloud Integration, Cloud Subscriptions, Cloud Migration, Cloud Service Dominance, Cloud Deployment, …]
- Cluster 6: [Social Media, User Generated Content, User Content, Personalized Content, Web Creators, Custom Models]
- Cluster 7: [AI Assistant, AI Search, AI Migration, Search Growth, AI Adoption, Artificial Intelligence, AI Marketing, AI Related Solutions]
- Cluster 8: [Cybersecurity Expansion, Cross Service Integration, Unified Experiences, Remote Productivity, Platform Expansion, Platform Shifts, Remote Work, Internet Growth, Platformization, Cyber Security, Collaboration Network, Agentforce Growth]
- Cluster 9: [Software Updates, Software Demand, CRM Disruption, Performance Tools, Broadband Necessity]
- Cluster 10: [Autonomous Database, Business Automation, Data Analytics, Data Companies, Data Extraction, Ecommerce Discovery]
- Cluster 11: [Mobile Apps, Video Monetization, Tech Innovation, Mobile Services, Mobile Offerings, Mobile Products]
- Cluster 12: [Automated Cropping, Video Features]
- Cluster 13: [Artificial Intelligence, Reasoning AI, AI Recommendation, AI Data Centers, AI Based, AI Data Center, Quantum Computing, Data Analytics, AI Infrastructure, Conversational AI, AI Innovation, 5g Ai Hpc, AI Adoption, Generative AI, AI Powered, Energy Efficient, Connected Computing, Autonomous Driving, …]
- Cluster 14: [Digital Media, Fixed Wireless, Materials Innovation, Digital Growth, High Bandwidth, Robotics Innovation, Digital Experience, HBM Capacity, HBM Adoption, Electro Optics, Digital Labor]
- Cluster 15: [Cloud Growth, Cloud Computing, Cloud Database, Cloud Based, Cloud Services, Cloud Adoption, Cloud Transformation, Cloud Guidance, Data Cloud, Cloud Security, Cloud Optimization, Cross Cloud, Cloud Infrastructure]
- Cluster 16: [Advanced Semiconductors, Chipset Advancements, Chip Innovation, Custom Accelerators, Advanced Packaging, SSD Adoption, Advanced SSDs, Custom CPUs, SSD Expansion, SSD Enterprise, Silicon Demand, Nextgen Chips, Custom ASIC, Silicon Photonics, Custom Silicon, Custom Nic, High Yield Chips, Optics Performance]
- Cluster 17: [Quantum Computing, Copackaged Optics, High Bandwidth, Edge Technologies, Edge Technology, Internet Things, N2 Technology, Edge Devices, Edge AI, Node Technology]
- Cluster 18: [Electrooptics Growth, Scale Up Fabrics, Wearables Growth, Data Center Growth, HBM Growth, Semiconductor Growth, Rack Scale Solutions, Rack Scale, System Level Scaling]
- Cluster 19: [Autonomous Vehicles, Advanced Cameras, AI Robotics, Robot Production, Robot Technology, Robotics Growth, Robot Delivery, Gate All Around, Lidar System]
- Cluster 20: [High Capacity DRAM, Low Power, Thin Film Battery, HBM Revenue, High Capacity Modules, High Performance SSD, Fast Memory, High Bandwidth Memory, Dram Shipments, Compute Memory, Full Rack Solutions, Nand Improvement, Euv Wafers]
- Cluster 21: [AI Smartphones, AI Glasses, AI Powered PCs, Industrial IoT, AI Appliances, AI Silicon, Custom AI Chips, AI Compute, AI Accelerators, Energy Efficient, Adaptive Computing, AI Powered Laptops, AI XPU, AI Servers, Smart Glasses, AI PCs, AI Custom Chips, Custom Logic]
- Cluster 22: [Compute Performance, HPC Demand, Next Gen GPUs, Nextgen GPUs, XPU Production, GPU Service, Next Gen XPU, Nextgen XPU, GPU Business, Mainframe Cycle, GPU Growth, GPU Adoption, Gaming CPUs, Compute Infrastructure, GPU Demand, GPU Deployment]
- Cluster 23: [Server Upgrades, Accelerated Infrastructure, Server Upgrade, Server CPU, Mac Upgrades, Device Upgrade]
- Cluster 24: [Mobile Storage, Mobile G9managed, Mobile Processors, Internet Devices, Premium Smartphone, Smartphone Replacement, Data Center, Device Capability, Device Innovation, Device Replacement, Mobile Offerings]
6. Conclusions
- Collecting relevant data (e.g., obtaining earnings calls transcripts for tech companies).
- Feeding data to an LLM that extracts relevant entity/relationship.
- Consolidating discovered entities.
- Storing extracted knowledge in a graph database.
- Querying the graph database to find the anchor entities.
- For each anchor entity, finding the corresponding numerical vector (by using Word2vec) of its label-type attribute. Clustering anchor entities using K-means and adaptive silhouette analysis.
- Running a top K filter on the clusters to isolate representative node–entity anchors.
- For each representative node–entity anchor, finding its neighboring within a specific max distance t and which are semantically close.
- Building the core document.
- Identifying the latent topics in the core document and finding trends.
- Given some brief context related to whatever subject and a user input search word, the LLM is asked to find similar and relevant other words (in this case, similar words are determined using a broader context that can be dynamically set). The input search word (or pattern), together with similar ones, will be used in a multiple-term regular expression search over the data stored in the knowledge graph.
- Once the outcome of a given search is returned by the graph database system, the results are clustered according to their source/date. After this step, a narrative corresponding to each cluster and based on the member entities/relationships names and properties might be built by the LLM.
- The clusters of trends obtained in the last step can be further processed by the LLM (for instance, a simple application might be the automatic cluster naming).
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
Appendix A. The Trends Identified for the Type Technological
- AI Initiatives, AI Revolution, Autonomy Advancement, AI Opportunity, Artificial Intelligence, AI Innovation, Product Innovation, AI Adoption, AI Leadership, Power Electronics, AI Growth, Silicon Photonics, AI Search, AI Models, Agentic AI, Small AI, Custom AI, Advanced IoT, IoT Advancements, Self Driving Cars, AI Instances, Open Source AI, Custom Hardware, AI Advancements, Autonomous Vehicles, 5G Adoption, AI Strategy, Edge Computing, AI Development, High-Speed Networking, AI Engineering, AI Expansion, AI Outcomes, Data Innovation, Hybrid Integration, Hardware Advancements, Autonomy Feature, AI Innovations, Platformization, AI Optimization, Transformational AI, Robotics Growth, High Performance Computing, AI Training, Data Analytics, AI Infrastructure, Advanced Process, Quantum Computing, Software Defined, AI Platform, Cloud Infrastructure, Cloud Adoption, AI Partnerships, AI Applications, AI Powered Devices, AI Data Center, AI Compute, AI Powered Tools, Cloud Expansion, AI Performance, AI Data Centers, AI Accelerators, AI Agent Collaboration, AI Collaboration, AI Capabilities, AI Strength, AI Notebooks, Hypergrowth AI, Multicloud Expansion, Semiconductor Growth, Cloud Growth, Advanced Packaging, AI Services, Innovative Technologies, Network Optimization, AI Solutions, Autonomy Features, Cloud Computing, Generative AI, AI Integration, AI Ecosystem, Cybersecurity Advancement, AI Market, AI Research, AI Personalization, XPU Production, Programmable Networks, Cybersecurity Demand, Cybersecurity Growth, Cloud Migration, Unified Data, AI Robotics, AI Momentum, AI Efficiency, Cloud Optimization, AI Scaling, Wireless Access, Silicon Software Integration, HBM Investment Increase, 5G Development, 5g Content, AI Offerings, 3D Dram, Hardware Renewal, Memory Innovations, Cloud Transformation, Mainframe Growth, AI Assist, Cloud Computing Decline, AI Demand Shift, Multicloud Adoption, AI Investments, Advanced Chip, Server Growth, Hardware Launch, AI Implementation, Digital Transformation, Smartphone Market, Advanced Memory, Advanced Logic, AI Insights, AI Workflows, AI Era, AI Guided Search, Optical Networking, AI Super Cycle, AI Orchestration, AI Workloads, NAND Improvement, Autonomy Opportunity, Silicon Design, IoT Growth, Language Models, AI Deployments, Compute Performance, Robotics Expansion, Robotics Adoption, Compute Memory Growth, GenAI Development, Copackaged Optics, Robotics Innovations, AI Productivity, Accelerated Computing, Autopilot Software, Real World AI, Node Growth, AI Progress, AI Control, AI Proliferation, AI Advances, AI Retail, Cyber Security, AI Transformation, Cloud Productivity, Multimodal Models, High Performance, AI Processing, Innovation Technologies, High Speed Products, Digital Upgrades, Cloud Performance, AI Automation, Automation Support, HBM4 Development, Virtual Assistant, AI Value, AI Chat, Cloud Integration, Cband Rollout, Linux Adoption, Cloud Automation, AI Deployment, Data Center Growth, Crosscloud Solutions, Autonomous Features, AI Products, Autonomous Transport, AI Assistant, AI Connect, Digital Economy, Digital Labor, AI CRM
Appendix B. The Trends Identified for the Type Digital
- Cloud Growth, Cloud Computing, Social Media, User Generated Content, User Content, Cloud Expansion, AI Assistant, AI Search, Cloud Transformation, Cybersecurity Expansion, Software Updates, Cloud Services, Cloud Transition, Cloud Native, Autonomous Database, Cloud Adoption, Cloud Infrastructure, Cloud Security, Software Demand, Cross Service Integration, CRM Disruption, Cloud Gaming, Cloud Automation, Business Automation, Mobile Apps, Performance Tools, Data Analytics, Automated Cropping, Cloud Integration, Cloud Subscriptions, Cloud Migration, Video Monetization, Unified Experiences, AI Migration, Personalized Content, Cloud Service Dominance, Cloud Deployment, Search Growth, Remote Productivity, Platform Expansion, Tech Innovation, AI Migrations, AI Adoption, Cloud Products, Artificial Intelligence, Data Companies, Platform Shifts, Remote Work, Internet Growth, Data Extraction, AI Marketing, Platformization, Cloud Consumption, Cyber Security, Hybrid Cloud, Video Features, Ecommerce Discovery, Cloud Environments, Mobile Services, Cloud Business, Web Creators, Mobile Offerings, Cloud Subscription, Mobile Products, Cloud Demand, AI Related Solutions, Collaboration Network, Cloud Commitments, Cloud Momentum, Broadband Necessity, Cloud Scaling, Cloud Efficiency, Custom Models, Agentforce Growth
Appendix C. The Trends Identified for the Type Technology
- Advanced Semiconductors, Chipset Advancements, Chip Innovation, Quantum Computing, Electrooptics Growth, Autonomous Vehicles, High Capacity DRAM, Low Power, AI Smartphones, AI Glasses, Scale Up Fabrics, Copackaged Optics, AI Powered PCs, Custom Accelerators, Compute Performance, Advanced Packaging, Advanced Cameras, High Bandwidth, Thin Film Battery, Industrial IoT, Wearables Growth, HPC Demand, AI Appliances, Edge Technologies, HBM Revenue, High Capacity Modules, Data Center Growth, SSD Adoption, High Performance SSD, AI Silicon, Custom AI Chips, AI Robotics, Edge Technology, Server Upgrades, Robot Production, AI Compute, AI Accelerators, Advanced SSDs, Custom CPUs, Accelerated Infrastructure, Fast Memory, Mobile Storage, Mobile G9managed, Robot Technology, Internet Things, SSD Expansion, SSD Enterprise, Mobile Processors, High Bandwidth Memory, Next Gen GPUs, Nextgen GPUs, Energy Efficient, Silicon Demand, N2 Technology, Adaptive Computing, Edge Devices, Nextgen Chips, Internet Devices, XPU Production, AI Powered Laptops, AI XPU, Robotics Growth, GPU Service, Next Gen XPU, Nextgen XPU, Dram Shipments, Server Upgrade, GPU Business, AI Servers, Custom ASIC, Mainframe Cycle, Premium Smartphone, Smartphone Replacement, Server CPU, Data Center, Smart Glasses, Robot Delivery, GPU Growth, Silicon Photonics, Gate All Around, HBM Growth, Semiconductor Growth, GPU Adoption, Custom Silicon, AI PCs, AI Custom Chips, Custom Nic, High Yield Chips, Custom Logic, Device Capability, Compute Memory, Gaming CPUs, Compute Infrastructure, Rack Scale Solutions, Rack Scale, Mac Upgrades, System Level Scaling, Full Rack Solutions, Optics Performance, HBM Technology, Lidar System, Nand Improvement, Device Innovation, GPU Demand, GPU Deployment, Edge AI, Device Upgrade, Device Replacement Hardware, Semiconductor Manufacturing, Node Technology, Euv Wafers, Mobile Offerings Hardware
Appendix D. The Trends Identified for the Type Innovative
- Artificial Intelligence, Custom AI, Digital Experience, Autonomy Progress, Autonomous Vehicles, Autonomous Assets, AI Powered, Custom Solutions Innovative Interconnect Technologies, Virtual Reality, Digital Transformation, Industrial IoT, Quantum Computing, Autonomous Cars, Data Analytics, Custom Products, Digital Labor, Edge Technologies, AI Access Capabilities, Generative AI, Metaverse Expansion, Autonomy Features, Edge Technology, Autonomous Driving, Next Gen AI, AI Ecosystem, AI Startups, Automation Future, Agent Automation, Silicon Integration, New Architecture Savings, Personalized AI, 3D Dram, Robotics Investment, Automation Adoption, Robotics Adoption, Gen AI, Frontier Models, Glasses Computing, Computer Vision, Optical Networking, Silicon Photonics, Augmented Reality, Agentic AI, Copackaged Optics, RandD Investment, Robotics Use, Gate All Around, M4 Based Products, Generative Audio, Generative Ai, Advanced Packaging, Custom Programs, Optical Neural, Mobileye Tech, AI Functionality, Open Source Databases, Autonomous Mode, Autonomous Transport, Autonomous Sales, Agentic Activity, AI Acceleration, Next Gen Products, Digital Technology
References
- Castro, A.; Pinto, J.; Reino, L.; Pipek, P.; Capinha, C. Large language models overcome the challenges of unstructured text data in ecology. Ecol. Inform. 2024, 82, 102742. [Google Scholar] [CrossRef]
- Perron, B.E.; Luan, H.; Victor, B.G.; Hiltz-Perron, O.; Ryan, J. Moving Beyond ChatGPT: Local Large Language Models (LLMs) and the Secure Analysis of Confidential Unstructured Text Data in Social Work Research. Res. Soc. Work. Pract. 2024; online first. [Google Scholar]
- Raja, H.; Munawar, A.; Mylonas, N.; Delsoz, M.; Madadi, Y.; Elahi, M.; Hassan, A.; Serhan, H.A.; Inam, O.; Hernandez, L.; et al. Automated Category and Trend Analysis of Scientific Articles on Ophthalmology Using Large Language Models: Development and Usability Study. JMIR Form. Res. 2024, 8, e52462. [Google Scholar] [CrossRef]
- Thapa, S.; Shiwakoti, S.; Shah, S.B.; Adhikari, S.; Veeramani, H.; Nasim, M.; Naseem, U. Large language models (LLM) in computational social science: Prospects, current state, and challenges. Soc. Netw. Anal. Min. 2025, 15, 4. [Google Scholar] [CrossRef]
- Tu, X.; He, Z.; Huang, Y.; Zhang, Z.; Yang, M.; Zhao, J. An overview of large AI models and their applications. Vis. Intell. 2024, 2, 34. [Google Scholar] [CrossRef]
- Du, K.; Zhao, Y.; Mao, R.; Xing, F.; Cambria, E. Natural language processing in finance: A survey. Inf. Fusion 2025, 115, 102755. [Google Scholar] [CrossRef]
- Kamath, U.; Keenan, K.; Somers, G.; Sorenson, S. Large Language Models: A Deep Dive; Springer: Cham, Switzerland, 2024. [Google Scholar]
- Atkinson-Abutridy, J. Large Language Models: Concepts, Techniques and Applications; CRC Press: Boca Raton, FL, USA, 2024. [Google Scholar]
- Amaratunga, T. Understanding Large Language Models; Apress: Berkeley, CA, USA, 2023. [Google Scholar]
- Geroimenko, V. The Essential Guide to Prompt Engineering: Key Principles, Techniques, Challenges, and Security Risks; Springer: Cham, Switzerland, 2025. [Google Scholar]
- Giray, L. Prompt engineering with ChatGPT: A guide for academic writers. Ann. Biomed. Eng. 2023, 51, 2629–2633. [Google Scholar] [CrossRef] [PubMed]
- Marvin, G.; Hellen, N.; Jjingo, D.; Nakatumba-Nabende, J. Prompt engineering in large language models. In Proceedings of the International Conference on Data Intelligence and Cognitive Informatics, Proceedings of the ICDICI 2023, Tirunelveli, India, 27–28 June 2023; Springer: Singapore, 2024; pp. 387–402. [Google Scholar]
- Pawlik, L. How the Choice of LLM and Prompt Engineering Affects Chatbot Effectiveness. Electronics 2025, 14, 888. [Google Scholar] [CrossRef]
- Jacobsen, L.J.; Weber, K.E. The Promises and Pitfalls of Large Language Models as Feedback Providers: A Study of Prompt Engineering and the Quality of AI-Driven Feedback. AI 2025, 6, 35. [Google Scholar] [CrossRef]
- Polat, F.; Tiddi, I.; Groth, P. Testing prompt engineering methods for knowledge extraction from text. Semant. Web 2025, 16, 1–34. [Google Scholar] [CrossRef]
- Han, B.; Susnjak, T.; Mathrani, A. Automating Systematic Literature Reviews with Retrieval-Augmented Generation: A Comprehensive Overview. Appl. Sci. 2024, 14, 9103. [Google Scholar] [CrossRef]
- Lakatos, R.; Pollner, P.; Hajdu, A.; Joó, T. Investigating the Performance of RAG and Domain-Specific Fine-Tuning for the Development of AI-Driven Knowledge-Based Systems. Mach. Learn. Knowl. Extr. 2025, 7, 15. [Google Scholar] [CrossRef]
- Guo, T.; Chen, X.; Wang, Y.; Chang, R.; Pei, S.; Chawla, N.V.; Wiest, O.; Zhang, X. Large Language Model Based Multi-agents: A Survey of Progress and Challenges. In Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence, International Joint Conferences on Artificial Intelligence Organization, Jeju, Republic of Korea, 3–9 August 2024; pp. 8048–8057. [Google Scholar]
- Wang, L.; Ma, C.; Feng, X.; Zhang, Z.; Yang, H.; Zhang, J.; Chen, Z.; Tang, J.; Chen, X.; Lin, Y.; et al. A survey on large language model based autonomous agents. Front. Comput. Sci. 2024, 18, 186345. [Google Scholar] [CrossRef]
- AI, ChatGPT, OpenAI. 2025. Available online: https://chatgpt.com (accessed on 25 May 2025).
- AI, Gemini, Google. 2025. Available online: https://gemini.google.com/ (accessed on 25 May 2025).
- AI, Copilot, Microsoft. 2025. Available online: https://copilot.microsoft.com (accessed on 25 May 2025).
- AI, Llama, Meta. 2025. Available online: https://www.llama.com (accessed on 25 May 2025).
- AI, Deepseek, High-Flyer. 2025. Available online: https://www.deepseek.com (accessed on 25 May 2025).
- Huggingface Leaderboard. 2025. Available online: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?ref=hub.athina.ai (accessed on 25 May 2025).
- Linz, P.; Rodger, S.H. An Introduction to Formal Languages and Automata; Jones & Bartlett Learning: Burlington, MA, USA, 2022. [Google Scholar]
- Radó, T. On non-computable functions. Bell Syst. Tech. J. 1962, 41, 877–884. [Google Scholar] [CrossRef]
- Dumitriu, D.C.; Sburlan, D.F. Enhanced Human-Machine Conversations by Long-Term Memory and LLMs. Int. J. User-Syst. Interact. 2023, 16, 85–102. [Google Scholar]
- Hornsteiner, M.; Kreussel, M.; Steindl, C.; Ebner, F.; Empl, P.; Schönig, S. Real-Time Text-to-Cypher Query Generation with Large Language Models for Graph Databases. Future Internet 2024, 16, 438. [Google Scholar] [CrossRef]
- Wang, Y.; Qin, J.; Wang, W. Efficient approximate entity matching using jaro–winkler distance. In Web Information Systems Engineering—WISE 2017, Proceedings of the 18th International Conference, Puschino, Russia, 7–11 October 2017; Springer International Publishing: Cham, Switzerland, 2017; pp. 231–239. [Google Scholar]
- Recchia, G.; Louwerse, M. A Comparison of String Similarity Measures for Toponym Matching. In Proceedings of the First ACM SIGSPATIAL International Workshop on Computational Models of Place (COMP ’13), Orlando, FL, USA, 5–8 November 2013; Association for Computing Machinery: New York, NY, USA, 2013; pp. 54–61. [Google Scholar]
- Liao, J.; Huang, Y.; Wang, H.; Li, M. Matching Ontologies with Word2Vec Model Based on Cosine Similarity. In Proceedings of the International Conference on Artificial Intelligence and Computer Vision (AICV2021). AICV 2021, Settat, Morocco, 28–30 June 2021; Hassanien, A.E., Haqiq, A., Tonellato, P.J., Bellatreche, L., Goundar, S., Azar, A.T., Sabir, E., Bouzidi, D., Eds.; Advances in Intelligent Systems and Computing. Springer: Cham, Switzerland, 2021; Volume 1377. [Google Scholar]
- DBpedia Spotlight. 2025. Available online: https://www.dbpedia-spotlight.org/ (accessed on 25 May 2025).
- Falcon 2.0. 2025. Available online: https://labs.tib.eu/falcon/falcon2/ (accessed on 25 May 2025).
- Tagme. 2025. Available online: https://tagme.d4science.org/ (accessed on 25 May 2025).
- Lu, Y.T.; Huo, Y. Financial Named Entity Recognition: How Far Can LLM Go? In Proceedings of the Joint Workshop of the 9th Financial Technology and Natural Language Processing (FinNLP), the 6th Financial Narrative Processing (FNP), and the 1st Workshop on Large Language Models for Finance and Legal (LLMFinLegal), Abu Dhabi, United Arab Emirates; 2025; pp. 164–168. [Google Scholar]
- Johnson, S.J.; Murty, M.R.; Navakanth, I. A detailed review on word embedding techniques with emphasis on word2vec. Multimed. Tools Appl. 2024, 83, 37979–38007. [Google Scholar] [CrossRef]
- Reimers, N.; Gurevych, I. Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, 3–7 November 2019; pp. 3982–3992. [Google Scholar]
- Ollama. 2025. Available online: https://ollama.com/ (accessed on 25 May 2025).
- Shorten, C.; Pierse, C.; Smith, T.; Cardenas, E.; Sharma, A.; Trengrove, J.; Luijt, B. StructuredRAG: JSON Response Formatting with Large Language Models. arXiv 2024, arXiv:2408.11061. [Google Scholar]
- Nananukul, N.; Sisaengsuwanchai, K.; Kejriwal, M. Cost-efficient prompt engineering for unsupervised entity resolution in the product matching domain. Discov. Artif. Intell. 2024, 4, 56. [Google Scholar] [CrossRef]
- Deloitte Tech Trends Report. 2025. Available online: https://www2.deloitte.com/us/en/insights/focus/tech-trends.html (accessed on 25 May 2025).
- Chang, Z.X.; Guo, W.; Wang, L.; Shao, H.Y.; Zhang, Y.R.; Liu, Z.H. Forecasting and analyzing technology development trends with self-attention and frequency enhanced LSTM. Adv. Eng. Inform. 2025, 64, 103093. [Google Scholar] [CrossRef]
- Promptfoo. 2025. Available online: https://github.com/promptfoo/promptfoo (accessed on 25 May 2025).
- PromptPerfect. 2025. Available online: https://promptperfect.jina.ai (accessed on 25 May 2025).
Chunk Size | No. of Words in Summary | No. of Chars | Total No. of Entities/Unique |
---|---|---|---|
400 | 979 | 5933 | 167/98 |
800 | 596 | 3538 | 79/45 |
Chunk Size | Total No. of Entities/Unique |
---|---|
400 | 124/72 |
800 | 70/31 |
Adobe (ADBE) | Booking Holdings (BKNG) | Meta Platforms (Facebook) | Qualcomm (QCOM) |
Adv. Micro Devices (AMD) | Broadcom (AVGO) | Microsoft (MSFT) | Salesforce (CRM) |
Alphabet (GOOG) | Cisco (CSCO) | MicroStrategy (MSTR) | ServiceNow (NOW) |
Amazon (AMZN) | CrowdStrike (CRWD) | NVIDIA (NVDA) | Super Micro Comp. (SMCI) |
Apple (AAPL) | Fortinet (FTNT) | NXPSemiconductors (NXPI) | TSMC (TSM) |
Applied Materials (AMAT) | Garmin (GRMN) | Oracle (ORCL) | Tesla (TSLA) |
AppLovin (APP) | IBM (IBM) | Palantir (PLTR) | Uber (UBER) |
ARM Holdings (ARM) | Intel (INTC) | PaloAltoNetworks (PANW) | Verizon (VZ) |
At&T (T) | Marvel Technolgy (MRVL) | Pinterest (PINS) |
Technological 450 | Digital 228 | Technology 153 | Hardware 145 | Financial 138 | Innovative 109 |
Software 102 | Innovation 75 | Artificial 73 | Security 73 | Economic 67 | Cybersecurity 62 |
Infrastructure 52 | Growth 43 | Transportation 41 | Industrial 34 | Computing 31 | Automotive 29 |
AI 29 | Marketing 25 | Adoption 21 | Telecom 20 | Sustainable 20 | Network 19 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Sburlan, D.F.; Sburlan, C.; Bobe, A. Tech Trend Analysis System: Using Large Language Models and Finite State Chain Machines. Electronics 2025, 14, 2191. https://doi.org/10.3390/electronics14112191
Sburlan DF, Sburlan C, Bobe A. Tech Trend Analysis System: Using Large Language Models and Finite State Chain Machines. Electronics. 2025; 14(11):2191. https://doi.org/10.3390/electronics14112191
Chicago/Turabian StyleSburlan, Dragoş Florin, Cristina Sburlan, and Alexandru Bobe. 2025. "Tech Trend Analysis System: Using Large Language Models and Finite State Chain Machines" Electronics 14, no. 11: 2191. https://doi.org/10.3390/electronics14112191
APA StyleSburlan, D. F., Sburlan, C., & Bobe, A. (2025). Tech Trend Analysis System: Using Large Language Models and Finite State Chain Machines. Electronics, 14(11), 2191. https://doi.org/10.3390/electronics14112191