Revolutionizing Social Robotics: A Cloud-Based Framework for Enhancing the Intelligence and Autonomy of Social Robots †

: Social robots have the potential to revolutionize the way we interact with technology, providing a wide range of services and applications in various domains, such as healthcare, education, and entertainment. However, most existing social robotics platforms are operated based on embedded computers, which limits the robot’s capabilities to access advanced AI-based platforms available online and which are required for sophisticated physical human–robot interactions (such as Google Cloud AI, Microsoft Azure Machine Learning, IBM Watson, ChatGPT, etc.). In this research project, we introduce a cloud-based framework that utilizes the beneﬁts of cloud computing and clustering to enhance the capabilities of social robots and overcome the limitations of current embedded platforms. The proposed framework was tested in different robots to assess the general feasibility of the solution, including a customized robot, “BuSaif”, and commercialized robots, “Husky”, “NAO”, and “Pepper”. Our ﬁndings suggest that the implementation of the proposed platform will result in more intelligent and autonomous social robots that can be utilized by a broader range of users, including those with less expertise. The present study introduces a novel methodology for augmenting the functionality of social robots, concurrently simplifying their utilization for non-experts. This approach has the potential to open up novel possibilities within the domain of social robotics.


Introduction
Social robots are increasingly being used in various domains, such as healthcare [1], education [2], business [3], telepresence [4], entertainment [5], and assistance [6]. They have the potential to provide social, behavioral, emotional, and cognitive support to people with diverse characteristics [7]. Furthermore, they can function as autonomous tools to support psychological health interventions [8]. The goal of social robots is to communicate and interact with humans in a manner that is socially acceptable and easily perceptible by humans [9]. In order to achieve their goals, social robots typically interact with several users for long periods of time [10]. In most applications of social robots, two primary constraints limit their ability to handle sophisticated human-robot interactions: one is related to the robot's hardware and software capabilities, and the other is related to its

Related Works
The field of cloud robotics has emerged as a way to enhance the capabilities of robots by utilizing the resources provided by cloud computing. This integration allows robots to access data, services, and other resources from remote locations, thus overcoming the limitations imposed by their onboard resources. Since the introduction of the "cloud robotics" term, various models for connecting robots to the cloud have been developed and investigated [23][24][25][26][27][28]. For instance, in [29], a customized monitoring interface for robots with a high degree of freedom was introduced. Those robots can be accessed through desktop and mobile devices. Their proposed interface utilizes standard web technologies and provides a 3D visualization of the robot's position and sensor data. The platform is user-friendly, with no software installation required, and it can be deployed on a cloud for scalability to support multiple viewers. The introduced platform has been successfully tested on multiple robots and browsers. Similarly, in [30], the author proposed a cloud robotics architecture for the Simultaneous Localization and Mapping (SLAM) problem by distributing computational tasks and data among remote servers, freeing robots from computational loads using the Hadoop framework and FastSLAM2.0. The proposed framework resulted in an accurate real-time performance for single-robot navigation.
Authors in [31] have introduced XBotCloud, a component of the XBot framework that allows robots to access cloud resources and perform tasks with real-time execution and communication performance. The proposed platform addresses security challenges and has been tested on multiple robotic platforms for performance and flexibility. In [32], the authors demonstrated the advantages of utilizing cloud-based applications in robotics, specifically for the purposes of face detection and identification. They introduced an autonomous platform-switching algorithm that combines a Parallax Arlo Robot System with cloud computing, leveraging the cloud as the primary source of computational power and energy. To optimize performance and enhance endurance while minimizing hardware requirements, they designed an execution controller scheme. The system features computer vision for human-target navigation and identification, as well as an interactive procedure for face-to-face interaction with the target user. The switching algorithm allows the robot to switch between available resources (robot and cloud VM) based on the network and battery statuses. In practical implementation, the introduced switching algorithm and cloud-aided system improve system performance and longevity. The RC-Cloud Robot System, as proposed in [33], is designed for cloud-based multitasking, such as real-time face detection and recognition, by connecting to cloud computing resources, big data, and sharing information among clients. The proposed framework also utilizes ROS to support distributed computation on two local computers. The experimental results demonstrate the efficient execution of computation-intensive tasks for robot clients. In [34], a low-cost robot system with an embedded Android phone in a ROS environment was proposed to reduce hardware complexity and improve reproducibility in robotics development, tested using two Android phones, a laptop, and a robot with an Arduino microcontroller. The results showed advantages in incorporating smartphone processing and sensing as well as heavy external processing.
Several research studies have also investigated modular frameworks for social robots to simplify the programming of high-level modules, reducing the workload for non-expert programmers [35]. AMIRO, introduced by [36,37], is a modular architecture consisting of independent components connected through a planning unit. The modularity of the system allows each module to be run individually, as well as easily be replaced. Each component must implement a specific set of methods corresponding to the tasks it can execute. The framework includes modules for navigation, vision, communication, and planning, and it can be expanded to include dialogue capabilities, ambient intelligence sensors, and a Beliefs Desire Intent architecture implemented over ROS for improved planning and acting capabilities. The framework was validated in lab settings using a Pepper robot. Coronado et al. [38] proposed a modular framework that moves the robotics architecture of social robots from easy-to-use to easy-to-develop, enabling novice users to create complex intelligent robot systems without coding. The framework was successfully validated on a NAO robot. The Robot Interfaces from Zero Experience (RIZE) framework [39] is a cross-platform system that allows for the generation of intelligent authoring behaviors for various commercialized robots. It is built on NodE Primitive NEP, which supports distributed and modular systems using different middlewares and communication patterns. RIZE uses a modular approach based on independent behaviors that can be reused in other RIZE-based programs.
Microsoft Azure has been utilized to power autonomous robots and AI applications, such as natural language processing, robot-powered process automation, and machine learning [40][41][42]. Therefore, the global cloud robotics market is predicted to grow significantly, reaching $157.8 billion by 2030 [43]. Furthermore, collaborations between companies such as Ansys and Microsoft have been established to integrate simulation tools, Azure cloud, HPC, and digital twin technologies in cloud engineering initiatives [44]. Based on those facts, the adoption of Azure offers several advantages for cloud robotics, including cost optimization, flexibility, scalability, enhanced security, simplified management and maintenance, high availability, and quality services with a unique design and a high level of privacy [45][46][47][48].
Examples of such implementation include "robotic quarterback", Davidovich Group's use of Azure 3D and AI in the cloud for inventory management, and ABB's implementation of Azure cloud technology to replace traditionally human-led roles [49][50][51]. In [52], for instance, the authors tested data linking the NAO robot and Microsoft Azure as IoRT network servers to enable two-way communication and found it to be feasible. On the other hand, in [53], the authors explored the concept of cloud-edge hybrid robotic systems for physical human-robot interaction using MS Azur.
The field of cloud robotics is still developing, and there is a lack of consensus on the tools and technologies used in creating these systems. Additionally, a user-friendly system remains a challenge, as many proposed platforms utilize advanced ROS systems, which may require a high degree of technical expertise for use and maintenance [52]. Moreover, there may be limitations in terms of scalability and compatibility with a wide range of robot types and their hardware and software configurations [53]. The present investigation proposes a simple cloud robotics framework that harnesses the advantages of cloud computing to augment the functionality of social robots. This framework aims to surmount the inherent limitations imposed by these robots' hardware and software while enhancing their ease of use.

Proposed Platform
In the development of our platform, we have established four primary tenets to guide its implementation. These include (1) ease of use for the operator; (2) modularity to facilitate parallel and incremental advancement of capabilities; (3) adaptability to integrate with a range of robots and cloud systems; and (4) interactivity and versatility. In the following sections, we provide in-depth each of these principles and how they were integrated into the system's design.

Ease of Use of Operators
Our platform uses a simple graphical user interface (GUI) to facilitate effective system operation by operators with limited programming knowledge, Figure 1. Our user-friendly GUI can reduce the system's complexity and provide an intuitive interface. The following highlights the functionality of the main keys on the GUI. • English chatbot: A chatbot tab that continuously recognizes speech from the microphone AI agent and synthesizes the chatbot response using the speaker AI agent in English. The backend of the GUI is crucial for guaranteeing the robustness and stability of the system, as it supports the various components and processes that make up the user interface, Figure 2. In our framework, the GUI's backend was powered by a suite of programming languages, including:

•
Text to speech: A feature allowing for the input of Arabic text from the operator, which is then synthesized into speech by the AI agent speaker. • Facial Expression: The capability to send commands to the facial expression AI agent to display emotions on the robot such as Happy, Sad, and Angry.

•
Speech to text (continuous): A testing tab that led to the concurrent and continuous speech recognition and dynamic display on the web for the "Chatbot tab". • Custom web run: A testing tab to ensure web operation.

•
Text chatbot: A testing tab for the chatbot, which takes operator text input and displays the chatbot response on the interface. • Chatbot speaker: A testing tab for the speaker AI agent and chatbot. • Chatbot: A tab that continuously recognizes speech from the microphone AI agent and synthesizes the chatbot response using the speaker AI agent in Arabic. • UAEU department of law: A test case for the robot to deliver information on the degree programs and other details for the UAEU Department of Law. The operator selects the desired information, and the robot then describes and clarifies it. • English chatbot: A chatbot tab that continuously recognizes speech from the microphone AI agent and synthesizes the chatbot response using the speaker AI agent in English.
The backend of the GUI is crucial for guaranteeing the robustness and stability of the system, as it supports the various components and processes that make up the user interface, Figure 2. In our framework, the GUI's backend was powered by a suite of programming languages, including: • PHP, which enabled server-side programming and facilitated functionality such as operator login/logout, admin sessions, access control, and task-loading screens. • MySQL, which enabled database querying and manipulation for the robot chatbot and the admin's user credentials. • Python, which provided standalone packages and AI agents specifically for the robot. • JavaScript (client-side), which enabled browser-side programming and dynamic browser display. • HTML, which was used for designing the user interface and theme.

•
Microsoft Windows Bat Script, which allowed for the execution of operating system commands and the concurrent launching of processes across different programming languages. • NODEJS, a JavaScript server that supported web sockets and the dynamic sending of data from other agents. • GO Lang, which was used as an experimental system-side concurrent processing language for the purpose of achieving faster execution speeds and improved performance.

Modularity for Parallel, Incremental Capability Growth
To build a robust framework, we have utilized the principles of the design science research process (DSRP) model [54]. The DSRP model was selected for its ability to offer a structured methodology for the design development and evaluation of design artifacts,

Modularity for Parallel, Incremental Capability Growth
To build a robust framework, we have utilized the principles of the design science research process (DSRP) model [54]. The DSRP model was selected for its ability to offer a structured methodology for the design development and evaluation of design artifacts, focusing on creating practical solutions to real-world issues. We have chosen the spiral development approach (SDP) for software development with a modular design for each prototype. The SDP involves iteratively creating prototypes with modular components to create flexible, scalable, and easily maintainable software. The offered framework aims to manage the modular functional cloud for robotics operations by treating each sensor as an individual AI agent and generating sub-AI agents as required for each task. The AI agents will communicate with the central "Meta AI agent," which will prioritize and make operational decisions, Figure 3. The framework design will mainly comprise the following components: Meta AI agent, Camera AI agent, GPS AI agent, Battery AI agent, Microphone AI agent, Speaker AI agent, and Navigation AI agent. Additional AI agents can be added as necessary to accommodate the specific needs of individual robots. For example, in the case of the BuSaif robot, two additional agents were required: A Facial Expression agent and a Conversation agent. to the cache memory and security. Measures such as encrypting credentials, hiding all executable and operating system codes from the web server, and adding a "secret" to executable codes that would only execute if the correct secret were taken. The third stage involved the integration of the server-side framework MDD (Model, Database, and Display). Procedural code was transformed into functional and object-oriented code. During the fourth stage, all robot operational tabs were transformed into the Goal design of multi-AI agents and a meta-AI. The Go language was adopted, and Go routine concurrent channels were utilized for the AI agent, instead of solely relying on the operating system concurrent and asynchronous background processes. Once the multi-AI agent design and Meta AI were in place, most of the processing was transferred to a Kubernetes cluster (K3S/K8S), a mini cloud with 32 ARM CPU Cores and 64 GB RAM. This cluster was used to scale up the robot's operational capacity while processing the AI agent actuation on the robot.   At the initial stage, our framework was designed to be operating-system agnostic; thus, it can be used with both Windows and Linux systems. This is achieved by using general operating system commands that can execute programming language scripts, as shown in Figure 4. The commands will run Windows Batch Script on Windows, shell or bash script on Linux, and Android terminal on Android. Each web platform will have a homepage for the robot and a login button for the robot's web control panel, allowing the operator to access and control the robot.
The evolved platform design involves constructing a web platform on a cloud device that is equipped with a public IP and has the ability to control public ports. This platform connects directly to a remote robot using socket technologies such as WebSockets and HTTP POST/GET protocols, facilitating the operations of the robot.
At the final stage, we outlined the development of a prototype for cloud and general design platforms for robots in a laboratory setting, with a specific focus on the cognitive assessment robot BuSaif. The spiral development life cycle was employed to guide the platform's development, and the following stages were undertaken. During the first stage, a different approach was employed for the general AI agent, with a unique design for each sensor. The operating system was utilized to launch processes concurrently and await commands through sockets or WebSockets. In the second stage, improvements were made to the cache memory and security. Measures such as encrypting credentials, hiding all executable and operating system codes from the web server, and adding a "secret" to executable codes that would only execute if the correct secret were taken. The third stage involved the integration of the server-side framework MDD (Model, Database, and Display). Procedural code was transformed into functional and object-oriented code. During the fourth stage, all robot operational tabs were transformed into the Goal design of multi-AI agents and a meta-AI. The Go language was adopted, and Go routine concurrent channels were utilized for the AI agent, instead of solely relying on the operating system concurrent and asynchronous background processes. Once the multi-AI agent design and Meta AI were in place, most of the processing was transferred to a Kubernetes cluster (K3S/K8S), a mini cloud with 32 ARM CPU Cores and 64 GB RAM. This cluster was used to scale up the robot's operational capacity while processing the AI agent actuation on the robot.

Adaptability to Robots and Cloud Systems
The introduced framework enables remote access to shared and dispersed computing resources over the internet. The platform is designed to be compatible with commonly used commercial cloud services providers such as Google Compute Engine and Microsoft Azure. Alongside the cloud infrastructure, the system utilizes a low-level component that can accommodate various robots, including the custom-made robot "BuSaif" as well as commercially available robots such as "Pepper," "NAO," and "Husky," all of which have undergone successful testing. The cloud infrastructure comprises high-performance servers, proxy servers, and databases, providing enhanced processing and storage capabilities. A diagram illustrating the architecture of cloud robotics systems and highlighting some of their applications is provided in Figure 5.

Adaptability to Robots and Cloud Systems
The introduced framework enables remote access to shared and dispersed computing resources over the internet. The platform is designed to be compatible with commonly used commercial cloud services providers such as Google Compute Engine and Microsoft Azure. Alongside the cloud infrastructure, the system utilizes a low-level component that can accommodate various robots, including the custom-made robot "BuSaif" as well as commercially available robots such as "Pepper," "NAO," and "Husky," all of which have undergone successful testing. The cloud infrastructure comprises high-performance servers, proxy servers, and databases, providing enhanced processing and storage capabilities. A diagram illustrating the architecture of cloud robotics systems and highlighting some of their applications is provided in Figure 5.

Interactivity and Versatility
The platform has been employed in three distinct scenarios. The first application involved the use of the BuSaif Robot in a cognitive assessment research project. The robot was equipped with various AI agents to detect and recognize individuals' faces and emotions, while also being capable of conducting communication sessions with both healthy

Interactivity and Versatility
The platform has been employed in three distinct scenarios. The first application involved the use of the BuSaif Robot in a cognitive assessment research project. The robot was equipped with various AI agents to detect and recognize individuals' faces and emotions, while also being capable of conducting communication sessions with both healthy individuals and those with cognitive impairments. Despite possessing only average programming skills, the operators were able to create the experimental scenario with ease. The second occasion involved the robot delivering opening speeches for three local conferences and interacting with guests by answering their questions about the event and program. The robot operators had no programming knowledge and were able to develop the scenario after a brief two-day training session. The last application involved the robot welcoming school students during their visit to the lab. In all of these instances, the BuSaif robot, owing to the developed platform, was able to leverage a broad spectrum of AI features, enabling it to effectively communicate with the public and exhibit advanced capabilities.
In order to ensure safe interactions with the robots, several standard safety features were considered. These features include the provision of an emergency stop button, collision detection and avoidance sensors, a safe movement speed, and secure data-handling mechanisms. By incorporating these safety features, operators and users can reduce the risk of accidents or injuries and ensure safe interactions with our social robots.
Our proposed cloud robotic platform incorporates all of the primary tenets discussed above and includes the following elements:

1.
Robots: Our platform supports a wide range of robots, including BuSaif, Pepper, NAO, and Husky.

2.
Cloud computing infrastructure: We utilize Google and Microsoft Azure as well as a database for our cloud computing needs.

3.
Communication networks: Our platform supports various communication networks, including sockets, web sockets, HTTP post and get, as well as peer-to-peer media streaming through AI agents such as Microphone and Camera.

4.
Control and management software: We use Meta-AI as our control and management software, as depicted in Figure 2.

5.
User interfaces: Our platform provides user interfaces, as illustrated in Figures 2 and 3.

Performance Analysis of a Cloud-Based Platform vs. a Local Platform for Robot Control
In this section, we present an analysis of the performance of the proposed cloud-based platform versus the local platform for controlling the BuSaif robot. The analysis includes the assessment of various aspects such as memory, speed, processors, and functionality. The originally used local PC, embedded with the BuSaif robot, served as the baseline for comparison. At stage 1 of the proposed design, the web platform was designed with a modular approach that divides each operation into a tab on the web platform control panel. Each operational tab is unique, and each tab launches multi-lingual programs concurrently and asynchronously via the operating system in the background. These operations communicate through sockets, web sockets, HTTP POST, and HTTP GET. The modular design of tabs simplifies the implementation, allowing for the ease of integrating new tabs for additional operations. The multi-processing design of the web platform enables the robot to scale up for heavy operations and ultimately communicate with the robot AI agents while processing all functions in a K3S/K8S cluster on the cloud. Table 1 demonstrates an overview of the structure comparison between the proposed and local platforms. Table 2 demonstrates an overview of the comparison results between the cloud-based and local platforms. The comparison revealed that the cloud-based platform outperformed the local platform in terms of speed and memory usage. The modular design of the proposed platform and the use of multi-lingual programs led to improved functionality and scalability, as evidenced by the ability to process complex operations and communicate with the robot AI agents. The proposed platform's multi-processing design also allows for distributed cloud computing, which is a significant advantage over the local platform's single-process approach.  This study provides evidence that a cloud-based robotics platform can be a viable alternative to a local platform for robot control, particularly in terms of speed, memory usage, functionality, and scalability. The modular design of the proposed platform and its multi-processing capabilities offer significant advantages over the local platform's single process approach. Future work should focus on optimizing the cloud platform further to improve its performance and usability.
The cloud-based framework has undergone testing on commercially available robots, namely the "Husky", "NAO", and "Pepper". Although a comparative analysis similar to Table 2 could not be conducted due to insufficient specification information, which depends on the robot version, it is worth noting that these robots can leverage the cloud to gain access to advanced features such as updated object detection models, facial expression analysis, cognitive capabilities [55], and the new chatGPT API for advanced natural language processing capabilities in both English and Arabic [56].

Responding Time
The initialization duration of the conventional Python graphical user interface (GUI), i.e., local PC, is approximately 9 s, with the initial 3 s utilized for the loading of the Python interpreter and its dependent classes, including Chatter, Face, Communicator, Eye, and Tracker. The remaining 6 s are mainly attributed to loading the TensorFlow and Caffee models. On the other hand, our cloud-based platform features a modular design that enables the authentication of administrators to the control panel in less than a second. The platform launches multi-lingual processes concurrently in the background per tab, ranging between 1 to 4 s, depending on the tab. The design objective is not solely to rely on multiprocessing and concurrency but also to avoid the use of interpreted languages by employing compiled languages such as C++ and Go, which provide fastercompiled machine code execution speeds instead of a slower interpretation during run time. Furthermore, Go, which is designed with concurrency in mind, is a suitable candidate for this web platform, given its support for goroutines and communication channels.

Microphone AI Agent
The implementation of the microphone AI agent involves utilizing Microsoft Azure cognitive services' speech SDK for speech recognition. The algorithm employed entails utilizing a loop to invoke the SDK to recognize speech from the microphone and transcribe it into Arabic text. The loop allows for up to 11 s for phrase recognition, with the process continuing until a phrase is recognized. Upon detection, speech recognition ceases, and the identified words are transmitted to the chatbot component of the system. To enhance the efficacy of the Microphone AI agent, the cloud platform launches a concurrent background process that leverages Microsoft cognitive services' continuous recognition function to attempt speech recognition every 0.5 s. The process is replicated four to six times, following which the predicted phrase recognition is relayed to the subsequent AI agent, i.e., the language and chatbot agent.
This algorithm operates continuously without halting, and it persistently recognizes and predicts phrases as an independent concurrent process.

AI Based Chatbot Agent
When utilizing the local PC on the robot, the classical Python GUI operates as a singular process that scans an intent folder containing labelled text files with phrases. It searches each labelled text file individually and calculates the probability of the phrase's occurrence for every file. The highest probability label is subsequently extracted from a response folder that also comprises labelled text files. From this folder, a phrase is randomly selected from a labelled text file and transmitted to the speaker agent. The presented platform follows the same principle but adopts a MySQL database for intent and response, which is populated with chatbot tables. The phrase in the cloud platform is conveyed from the microphone AI agent as an HTTP post to the web server, which searches the intent database for the phrase. If the phrase is detected, a random row is selected from the labelled response database and transmitted to the speaker AI agent for speech synthesis.
The updated platform features a database in a K3S/K8S Kubernetes cluster, with the phrase transmitted to the master node. The master node initiates a concurrent Go language routine, which divides the search. For instance, if there are 1000 tables in the intent database, the master node creates 20 concurrent Go routines, with each responsible for searching 50 tables. The master node then sends these requests to the chatbot service node port, distributing the load between the cloud processing units. The phrase is subsequently conveyed via socket and port to the next AI agent, i.e., the speaker's AI agent.

Speaker AI Agent
The speaker AI agent employs the Microsoft cognitive services speech SDK to convert Arabic text phrases into sound for the speaker, utilizing the text-to-speech technology. Both the Python GUI and the web platform leverage the Microsoft speech synthesis function. The sole distinction is that the web platform already launches the speaker AI agent as a concurrent background process that listens on a socket and port to synthesize any message into speech. The advanced design follows the same principles as the cloud platform, but it utilizes a compiled language to boost the response speed.

Camera AI Agent
The classical Camera AI agent employs the Python TensorFlow framework and Caffee model for object detection. Nevertheless, it is linked directly to the robot's operations owing to its high processing requirements. In contrast, the cloud platform's Camera AI agent is reliant on peer-to-peer media stream connection between the robot and operator's browser. It also utilizes TensorFlow JavaScript to load the model, which provides superior image processing speed.

Discussion
The current limitations of social robots can be attributed to two primary factors: limited computational power and difficulty to setup and operate for non-expert programmers. Due to their limited computational power, most existing robots are unable to perform complex algorithms, respond to multiple inputs, or understand and process complex hu-man language or emotions. These limitations also impact the robot's ability to learn and adapt to new situations. For example, the NAO robot has a maximum of 2 GB of internal memory and a 1.6 GHz Intel Atom processor, which limits its ability to recognize advanced facial and speech expressions in real-time or perform trust inference based on unstructured data. To address these limitations, the utilization of cloud resources has been suggested as a potential solution, which facilitates the delegation of computational responsibilities to external servers by the robots. Consequently, this strategy can augment the robots' cognitive abilities, enabling them to conduct customized interactions in accordance with the individual users and environmental conditions. The cloud can enable robots to perform more challenging tasks such as object recognition and manipulation, social navigation, and advanced human-robot interactions. In addition to limited computational power, setting up and programming existing robots can also be challenging for non-expert programmers. The Robot Operating System (ROS), which is widely used to control robots, can be difficult to learn and use for novice users with little to no experience [57]. Real-time performance and setting up most robotic manipulation systems are also challenging that need to be addressed.
As technology continues to evolve, the limitations of social robots are expected to reduce, and a cloud robotic platform is proposed in this study that aims to increase the acceptance of robotic services by non-expert programmers. The proposed platform has the potential to facilitate the development of socially intelligent and user-friendly robots that can overcome the limitations associated with traditional stand-alone robots, which are difficult to program and have limited computational power. By leveraging cloud resources, the proposed platform aims to enable the creation of robots that are not only smarter but also more accessible to the public.
The cloud-robotics paradigm represents a promising approach for enhancing the efficiency, cost-effectiveness, and intelligence of robots via cloud computing capabilities [57]. However, the absence of a platform that can guarantee compatibility with a diverse range of robotics platforms and provide ease-of-use for non-expert users remains a challenge. Despite the emergence of various open-source and proprietary options, the research field of cloud robotics is still evolving, leaving ample opportunities for further exploration. In this study, we introduce a Meta AI cloud platform that is readily adoptable by many common social robots and is user-friendly for non-experts. Nevertheless, the platform's simplicity may limit its ability to handle complex robots, a trade-off that we deem acceptable given our target user base. Achieving both the handling of complex robots and non-expert usability is a challenging task that needs more explorations.

Conclusions
Social robots have the potential to revolutionize the way we interact with technology, providing a wide range of services and applications in various domains. However, most of the social robotics platforms are limited by their reliance on embedded computers, which limits their access to advanced AI-based services and online platforms. This study aims to enhance the capabilities of social robots by proposing a platform that utilizes the benefits of cloud computing and clustering. Our proposed platform was tested on a customized and commercialized robot, and the results indicate that the platform can handle both kinds of robots and allows for a user-friendly interface for general users. The implementation of our proposed platform has the potential to result in more intelligent and autonomous social robots that can be utilized by a wider range of users, including those with less expertise.