Next Article in Journal
Topological Overview of Auxiliary Source Circuits for Grid-Tied Converters
Next Article in Special Issue
Editorial: Social Manufacturing on Industrial Internet
Previous Article in Journal
A Crack Defect Detection and Segmentation Method That Incorporates Attention Mechanism and Dimensional Decoupling
Previous Article in Special Issue
Research on the High Precision Synchronous Control Method of the Fieldbus Control System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Efficient Product-Customization Framework Based on Multimodal Data under the Social Manufacturing Paradigm

1
State Key Laboratory for Management and Control of Complex Systems, Beijing Engineering Research Center of Intelligent Systems and Technology, Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China
2
School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing 100049, China
3
Intelligent Manufacturing Center, Qingdao Academy of Intelligent Industries, Qingdao 266109, China
4
Guangdong Engineering Research Center of 3D Printing and Intelligent Manufacturing, The Cloud Computing Center, Chinese Academy of Sciences, Dongguan 523808, China
*
Author to whom correspondence should be addressed.
Machines 2023, 11(2), 170; https://doi.org/10.3390/machines11020170
Submission received: 30 November 2022 / Revised: 31 December 2022 / Accepted: 2 January 2023 / Published: 26 January 2023
(This article belongs to the Special Issue Social Manufacturing on Industrial Internet)

Abstract

:
With improvements in social productivity and technology, along with the popularity of the Internet, consumer demands are becoming increasingly personalized and diversified, promoting the transformation from mass customization to social manufacturing (SM). How to achieve efficient product customization remains a challenge. Massive multi-modal data, such as text and images, are generated during the manufacturing process. Based on the data, we can use large-scale pre-trained deep learning models and neural radiation field (NeRF) techniques to generate user-friendly 3D contents for 3D Printing. Furthermore, by the cloud computing technology, we can achieve more efficient SM operations. In this paper, we propose an efficient product-customization framework that can provide new ideas for the design, implementation, and optimization of collaborative production, and can provide insights for the upgrading of manufacturing industries.

1. Introduction

With the rapid development of social computing and social manufacturing (SM), traditional manufacturing mode is transforming to customization, small batch, digital production [1]. The cost of rapid prototyping equipment, represented by 3D printers, is decreasing, and the technology is becoming increasingly popular. Desktop and industrial-grade 3D printers make it possible for individuals to produce products independently. Users can personalize and manufacture physical objects in small batches without relying on complex factories as long as they have valid 3D model data (or have other types of manufacturing instructions). Each 3D printer is an independent manufacturing node with a naturally distributed, decentralized character. Anyone with an idea for a product design, even if they do not have the knowledge of 3D modeling, can design a digital model by relying on AI technology, and then realize it with manufacturing facilities. With a well-developed design platform, innovative ideas can be realized on the Internet. Using distributed manufacturing tools such as 3D printers, personalized and customized products can be provided to users worldwide.
Industry 4.0 utilizes the cyber-physical system (CPS) in combination with additive manufacturing, the Internet of things, machine learning, and artificial intelligence technologies to a realize and smarten manufacturing [2]. It aims to improve the intelligence level of the manufacturing industry and establish a smart manufacturing model with adaptability and resource efficiency. Cyber-physical-social systems (CPSS) further enhance the integration of Internet information and social manufacturing systems and fully integrate industry with the human society [3]. In the CPSS industrial environment, social manufacturing seamlessly connects the Internet of Things with 3D printers. The public can fully participate in the entire lifecycle of the product manufacturing process, to achieve real-time, personalized, large-scale innovation, and “agile mobile intelligence”.
Social manufacturing is a paradigm that uses 3D printing to involve the community in the process of producing products. During the 3D printing process, the first step is to obtain a 3D model. Currently, most 3D models are generated by computer-aided design (CAD), but the time and human resources costs are high. It is essential to design a user-friendly end-to-end 3D content generation model so that users can personalize the 3D model using text representation. Futhermore, a cloud platform can be built and put it into practical use.
The rest of the paper is organized as follows. In Section 2, we present the literature review of SM. Section 3 introduces the theoretical concepts. Section 4 reviews the current difficulties in product customization. Finally, in Section 5, the multimodal data-based product customization manufacturing framework is proposed and illustrated with practical application examples, followed by concluding remarks in Section 6.

2. Literature Review

Utilizing SM is a practical way to achieve intelligent, social, and personalized manufacturing. SM allows for the active participation of all facets of social customization and the intense personalization of every component of every product. The term SM was proposed by Professor Fei-Yue Wang at the “Workshop on Social Computing and Computational Social Studies” in 2010 [3], and its formal definition is given in his article in 2012 [4]: “The social manufacturing refers to the personalized, real-time and economic production and consumption mode for which a consumer can fully participate in the whole lifetime of manufacturing in a crowdsourcing fashion by utilizing the technologies, such as 3D printing, network, and social media.” Thus, it is noted that SM is inspired by the social computing concept that has been hailed as a cutting-edge manufacturing solution for the coming era of personalized customization. In the same way that the social computing [5] enables everyone in the society to obtain computing capacity, the SM enables everyone to obtain innovation and manufacturing capacity.
In the context of the SM, the critical feature is that everyone participates in product design, manufacturing and consumption in social production. Everyone can turn their ideas into reality. SM is a further the development and continuation of the existing crowdsourcing model. Xiong et al. [6] summarize this feature as “from mind to product”. Based on this concept, Shang et al. [7,8,9] provided a vision of connected manufacturing services that are smart and engaging online. Mohajeri et al. [10,11,12] argued for using social computing and other Internet technologies to realize personalized, real-time, and socialized production. Jiang et al. [13,14,15,16] focused on the mechanisms for crowdsourcing and outsourcing services throughout the entire lifespan of individualized production. As for the SM service mode, Xiong et al. [17] presented a social value chain system that applies the SM mode to the entire value chain and makes a contribution to bringing more potential to value-adding for all involved participants while reducing waste. Cao et al. [18] provided a manufacturing-capability estimation model for the SM mode’s service level that is based on the ontology, the semantic web, the rough set theory, and a neural network. It includes models for production service capability and machining service capability. In the evaluation system proposed by Xiong et al. [19], the best supplier is selected by using the hierarchical analysis method (AHP) and fuzzy comprehensive method. The suggested techniques can assist prosumers in assessing their suitability for a certain manufacturing requirement. Additionally, it can assist prosumers in their search and ability matching. Ming et al. [20] proposed a data-driven view in the context of SM. Xiong et al. [21] summarized the five layers that make up an SM system architecture, and a detailed discussion of each of the layers is presented next.
(1) Resource Layer: The production and service resources found in this layer can be combined to form a social resource network. A few of the resources include 3D printers, sophisticated logistics networks, information networks using 3D intelligent terminals, and heterogeneous operating systems. In this layer, 3D intelligent terminals can offer interaction and perception capabilities, and a modern logistics network can offer the logistical capability. Along with a range of physical links, the logistics network offers transparent information transmission options. For terminal perception, information communication, and production and product logistics, this layer is essential.
(2) Support Layer: Service module encapsulation, service discovery, service registration, service management, service data management, and middleware management are all part of this layer. Enterprise information can be transferred reliably, efficiently, and safely in accordance with user needs using the service registration and service discovery modules.
(3) Environment Layer: This layer includes the computer environment, information analysis environment and monitoring management environment that works in coordinated operation.
(4) Application Layer: This layer includes 3D modeling, operational monitoring tools, a platform for collaborative manufacturing management, a data collection and analysis platform, an optimization tool for allocating socialized manufacturing resources (SMRs), an assessment mechanism for system management, real-time monitoring, and resource scheduling.
(5) User Layer: This layer is composed of manufacturers, prosumers, and SMRs. Crowdsourcing is used to connect manufacturers and prosumers, and SMRs can be tailored for production.

3. Concepts

3.1. 3D Modeling

Additive manufacturing (AM) is a fabrication method that uses a digital prototype file as the foundation for constructing objects by printing layer by layer using material, such as powdered metal or plastic. Without the need for traditional tools, fixtures, or multiple machining processes, this technology enables the rapid and precise manufacturing of parts with a complex shape in a single machine, thereby allowing “free-form” manufacturing. Moreover, the more complex the structure of the product, the more significant the accelerating effect of its manufacturing. Currently, AM technology has many applications in the aerospace, medicine, transportation and energy, civil engineering, and other fields. With the development of AM technology and the popularity of 3D printers, there is an increasing demand for 3D modeling. The traditional modeling approach uses forward modeling software—3DMax, AutoCAD, etc. [22]—which is highly technical. It takes a long time to design a model, making it difficult to build complex, arbitrarily shaped objects. Roberts et al. [23] proposed a method to obtain 3D information from 2D images, known as reverse modeling. Since then, vision-based 3D reconstruction developed rapidly, and many new methods have emerged. Image-based rendering (IBR) methods could be divided into two categories depending on whether they rely on a geometric prior or not. Those methods that rely on geometric priors generally require a multi-view stereo (MVS) algorithm first to calculate the geometric information of the scene, and then to guide the input image for view composition. The methods that do not rely on geometric priors can generate new views directly from the input images. In 2018, Yao proposed an unstructured multi-view 3D reconstruction network (MVSNet) [24], an end-to-end framework using deep neural networks (DNNs). IBR methods that rely on geometric a priori information perform better when geometric information is abundant and accurate. Still, when the geometry is missing or incorrect, it produces artifacts and degrades the quality of the view. The light field is the main area of research that does not rely on geometric a priori methods. Traditional methods for drawing light fields require dense and regular image captures, making them difficult to apply in practice. In recent years, with the development of DNN, researchers have discovered that it is possible to synthesize new views by fitting a neural network to a light sample of the scene, thereby implicitly encoding the light field of the input image. Compared to traditional light fields, the neural reflectance field (NeRF) [25] method can be used for handheld captures with a small number of input images, greatly extending the applicability. The main advantages and disadvantages of the mainstream 3D modeling approach are shown in Table 1. As shown in Figure 1, NeRF can be briefly summarized as using a multi-layer preceptron (MLP) neural network that learns a static 3D scene implicitly. While the input is the spatial coordinates, the view direction output is the bulk density of the spatial location under that view direction and the view-related camera light radiation field. In its basic form, a NeRF model represents scenes as a radiance field approximated by a neural network. The radiance field describes color and volume density for every point and for every viewing direction in the scene. This is written as,
F ( x , θ , ϕ ) ( c , σ )
where x = ( x , y , z ) are the in-scene coordinates, ( θ , ϕ ) represent the azimuthal and polar viewing angles, c = ( r , g , b ) represents color, and σ represents the volume density. This 5D function is approximated by one or more MLP, sometimes denoted as F. The two viewing angles ( θ , ϕ ) are often represented by d = ( d x , d y , d z ) , a 3D Cartesian unit vector. This neural network representation is constrained to be multi-view consistent by restricting the prediction of σ , the volume density (i.e., the content of the scene), to be independent of viewing direction, whereas the color c is allowed to depend on both viewing direction and in-scene coordinates. NeRF utilizes a neural network as an implicit representation of a 3D scene instead of the traditional explicit modeling of point clouds, meshes, voxels, etc. Through such a network, it is possible to directly render a projection image from any angle at any location. For this purpose, NeRF introduces the concept of radiation field, which is a very important concept in graphics, and here we give the definition of the rendering equation,
L o ( x , d ) = L e ( x , d ) + Ω f r x , d , ω i L i x , ω i cos θ d ω i
The neural radiation field represents the scene as volumetric density and directional radiation brightness at any point in space. Using the principles of classical stereoscopic rendering, we can render the color of any ray passing through the scene. The bulk density σ ( x ) can be interpreted as the derivable probability that the ray stays at position X for an infinitesimal particle. Under the conditions of the nearest boundaries t n and farthest boundaries t f , the color C ( r ) of the camera light is:
C ( r ) = t n t f T ( t ) σ ( r ( t ) ) c ( r ( t ) , d ) d t , where T ( t ) = exp t n t σ ( r ( s ) ) d s
In practical applications, a fully automated, full-flow 3D reconstruction tool chain is urgently needed in many areas across industries, which also opens up many ideas and opportunities for our 3D reconstruction research. There are two key issues in this process that deserve our consideration. The first is that in the construction process of a 3D reconstruction system, the traditional geometric vision has clear interpretability, so how can we integrate an end-to-end deep learning method? Whether it is an end-to-end replacement or one embedded in our current process seems to be inconclusive and needs further exploration. The second is, from this practicality, how can we combine the reverse reconstruction in computer vision with the forward reconstruction in graphics so as to truly realize the highly structured and highly semantic 3D model required by industry from massive images. Generation is also an essential trend in the future. Only by solving these problems can we indeed use our image 3D reconstruction system to effectively support various business requirements in practical application settings, such as digital city planning, VR content production, and high-precision mapping.

3.2. Cloud Manufacturing

Cloud manufacturing is an intelligent, efficient, and service-oriented new manufacturing model proposed in recent years. We aim to organically combine the 3D content generation model and cloud manufacturing mode so that this technology can be successfully implemented and applied to actual production. Cloud manufacturing is a cutting-edge manufacturing model that combines cloud computing, big data, and the Internet of Things to enable service-oriented, networked, and intelligent manufacturing. It provides a customized approach to production based on the use of widely distributed, on-demand manufacturing services to meet dynamic and diverse individualized needs and to support socialized production [26,27,28,29]. Cloud manufacturing emphasizes the embedding of computing resources, capabilities, and knowledge into networks and environments, allowing the center of attention of manufacturing companies to shift or return to users’ needs themselves. Cloud manufacturing is committed to building a communal manufacturing environment where manufacturing companies, customers, intermediaries, and others can fully communicate. In the cloud manufacturing model, user involvement is not limited to the traditional user requirement formulation and user evaluation but permeates every aspect of the entire manufacturing lifecycle. In the cloud manufacturing model, the identity of the customer or user is not unique, i.e., a user is a consumer of cloud services but also a provider or developer of cloud services, reflecting a kind of user participation in manufacturing [30]. The advanced 3D-printing cloud model integrates 3D printing technology with the cloud manufacturing paradigm. Furthermore, the advanced 3D-printing cloud-model-based 3D-printing cloud platform has a service-oriented architecture, personalized customization technology, and a scalable service platform [31]. The 3D-printing cloud platform is not a standalone information-sharing platform, but rather an open, shared, and scalable platform that offers both 3D printing and other types of high-value-added knowledge services. The 3D-printing cloud platform can power a 3D printing community based on collaborative innovation and promote the growth of 3D-printing-related industries. Manufacturing resources (such as 3D printers and robots) from various regions and enterprises are encapsulated into various types of manufacturing services in the cloud platform, with the goal of providing service demanders with on-demand service compositions. Yang et al. [32] provided a cloud-edge cooperation mechanism for cloud manufacturing to offer customers on-demand manufacturing services, greatly enhancing the usage of distributed manufacturing resources and the responsiveness to the needs of customized products. Tamir et al. [33] proposed a new robot-assisted AM and control system framework which effectively combines 3D printing and robotic arm control to better support cloud printing tasks.

4. The Difficulties in Product Customization

Offering products with their colors, components, and features changed to suit consumers’ preferences is known as product customization. The practical implementations of NeRF-based product customization currently face several difficulties due to the limitations of technology and resources. This section evaluates the research and divides the challenges into three groups,
(1) The accuracy of 3D modeling: We need to understand and render 3D scenes, because we human beings already live in such a 3D world, and we need to interact with others. When making a virtual scene, we hope to reconstruct the object from different perspectives and then analyze and observe the object. In the medical field, we hope to reconstruct the parts of each person and then guide the doctor’s decisions. At the same time, we hope to have a seamless interaction with this virtual world, hoping to get realistic enjoyment in the virtual world. For the next generation of artificial intelligence, we hope that it can understand 3D scenes so that it can better serve humans. Hopefully, we can make artificial intelligence have the ability to interact. This will form a pipeline. After reconstructing the 3D representation from the 2D world, a realistic rendering can be made. On top of this, it is hoped that a 3D scene could be generated so that the generated model could be used to learn the entire process. However, there are specific difficulties in realizing this. In fact, the generation of 3D scenes is a very complicated process. Take NeRF as an example, it requires multi-view images. However, after all, the amount of data in multi-view images is far from enough. This kind of data is actually a lot of this single-view images on the Internet, which lacks perspective information. We hope to use the existing pre-trained model to provide prior knowledge.
(2) Security and supervision: AM files can be easily transferred from the AM design stage to the shop floor during final production. Digital supply networks and chains can be created thanks to the ease with which parts and products can be shared and communicated thanks to AM’s digital nature. In addition, these digital advantages come with a few drawbacks. A digital design-and-manufacturing process increases the likelihood of data theft or tampering without a robust data protection framework. Data leakage and identity theft will pose significant security risks as the scale of the SM production system grows, and the participation of dishonest and malicious nodes will put the interests of honest nodes in jeopardy. At the same time, when physical products are transported via product data, it is essential to secure, store, and share data containing all important information. The digital thread for AM, also known as DTAM, creates a single, seamless data link between initial design concepts and finished products in order to mitigate risks. The integration of multiple printers and printing technologies and a number of distinct and disjointed physical manufacturing facilities is the primary obstacle in this challenge. Additionally, because parts must be inspected throughout the process rather than just at the end, businesses can need help to keep track of events that take place during the additive process. This may be necessary for part certification and qualification.
(3) Production efficiency: The uneven distribution of manufacturing resources in the manufacturing industry and the low utilization rate of resources have seriously affected the development of the manufacturing industry. To effectively integrate scattered manufacturing resources and improve the utilization rate of manufacturing resources, cloud manufacturing, a manufacturing model that uses information technology for services, appears in everyone’s vision. As one of the key research issues of the cloud manufacturing platform, the scheduling of manufacturing resources in the cloud manufacturing environment will affect the overall operational efficiency of the cloud manufacturing platform. In both academia and industry, cloud computing resource scheduling problems are considered to be as difficult as non-deterministic polynomial optimization problems, that is, NP problems. Therefore, algorithms that solve relatively conventional scheduling problems may suffer from dimensional damage when the scale of the problem increases. With the development of cloud computing and the increase in complexity, this problem has become more challenging.

5. The Efficient Product Customization Framework

Considering that the characteristics of NeRF and other artificial intelligence technology have great potential for solving the difficulties of SM, a customization production framework of 3D printing using NeRF and ultra-large-scale pre-trained multimodal models as the baseline is proposed in this paper. As shown in Figure 2, the framework is constructed in the order of the multimodal data-based customization production process and value flow, specifically, from the bottom to the top, it is divided into three parts: 3D modeling service, blockChain encryption service, and cloud management service.

5.1. 3D Modeling Service

3D content customization is a very challenging task as it is a 3D representation. The existing stylized methods are 2D, so we can use the 2D stylized form to provide a kind of supervision. Still, this kind of supervision does not have 3D information, so we need to use the mutual learning between 2D and 3D. Two-dimensional neural networks can provide a stylized reference. More importantly, we offer this 3D-based spatial consistency information to 3D NeRF and finally stylize NeRF. Therefore, we can convert the scenes taken by mobile phones into this 3D stereo-stylized effect.
To solve the difficulties in 3D modeling, as shown in Figure 3, the service includes two modules: image-based 3D digital-asset reconstruction and text-based 3D digital-asset generation. In the image-based module, the end-to-end model can automatically generate a 3D geometric model without human intervention by taking multiple 2D pictures of the object from different angles with a device such as a mobile phone. The entire model is implemented in the following steps. First, a 360° surrounding image needs to be captured at a fixed focal length. After that, we use this tool for pose estimation. After recovering the poses of the photographed objects, the poses obtained from the sparse reconstruction are converted into local light field fusion (LLFF)-format data and then rendered by NeRF. The NeRF combines light-field sampling theory with neural networks, using an MLP to implicitly learn the scene in the sampling of all light, and it achieves good view synthesis results. In addition, NeRF uses the images themselves for self-supervised learning, which is applicable to a wide range of datasets, from synthetic to real-world. Finally, the corresponding mesh model is exported. To make NeRF practical for 3D modeling in AM, we build it based on instant neural graphics primitives (Instant-NGP) [34], a fast variant of NeRF. In addition, in order to achieve the realization of the reconstruction process, only the target object is modeled. We use image segmentation algorithms in the pre-processing process, which focuses on the accurate extraction of the reconstructed target and can reduce redundant information in the image during the reconstruction process.
The process of 3D reconstruction based on sequenced images is rather complex. In order to further enhance the convenience and versatility of product customization, we propose a text-guided 3D digital-asset generation module. Recent breakthroughs in text-to-image synthesis have been driven by multimodal models trained on billions of image-to-text pairs. The CLIP multimodal pre-training model is called contrastive language-image pre-training [35], i.e., a pre-training method based on contrasting text-image pairs. CLIP uses text as a supervised signal to train the visual model, which results in a very good zero-shot effect and good generalization of the final model. The training process is as follows: The input of CLIP is a pair of picture–text pairs. The text and images are output with corresponding features by the text encoder and image encoder, respectively. The text features and image features are then compared and learned on these outputs. If the input to the model is n pairs of image–text pairs, then this pair of mutually paired image–text pairs are positive samples (the parts marked in blue on the diagonal of the output feature matrix in the Figure 4), and the other pairs are negative samples. The training process of the model is thus to maximize the similarity of the positive samples and minimize the similarity of the negative samples. However, applying this approach to 3D synthesis requires large-scale datasets of labeled 3D assets and efficient methods for denoising 3D data.
The traditional NeRF scheme for generating 3D scenes requires multiple 3D photographs to achieve 360° visual reconstruction. In contrast, the Dreamfields [36] algorithm used in this paper does not require photos to generate 3D models and can generate entirely new 3D content. In fact, the algorithm is guided by a deep neural network (DNN) that can display geometric and color information based on the user’s textual description of the 3D object and some simple adjustments. When training the algorithm, a multi-angle 2D photo is required as a supervised signal, and once trained, a 3D model is generated and a new view is synthesized. The role of the CLIP multimodal pre-training model is to evaluate the accuracy of the text-generated images. After the text is fed into the network, the untrained NeRF model generates random views from a single perspective, and then the CLIP model is used to evaluate the accuracy of the generated images. In other words, as shown in Figure 3, NeRF renders the image from a random position, and finally uses CLIP as a measure the similarity between the text description and the composite image of given parameters θ and position s orientation p. This process is repeated 20,000 times from different views until a 3D model is generated that matches the text description. The corresponding loss function for this training process is:
L CLIP ( θ , pose p , caption y ) = g ( I ( θ , p ) ) T h ( y )
L T = min ( τ , mean ( T ( θ , p ) ) )
L total = L CLIP + λ L T
where g ( · ) is the image encoder and h ( · ) is the text encoder. The L c l i p objective is to maximize the cosine similarity of the rendered image embedding to the text embedding, and the L t objective is to maximize the average transmittance. Inspired by Dreamfields algorithm, this paper leverages a priori knowledge from large pre-trained models to generate 3D digital assets better. Using Dreamfields as the baseline, the images are encoded by a CLIP image encoder and compared with the text input encoded by a CLIP text encoder. To implement support for Chinese prompts, we replaced CLIP with Taiyi-CLIP [37], a visual-language model using Chinese-Roberta-wwm [38] as the language encoder, and applied the vision transformer (ViT) [39] in CLIP as the vision encoder. They froze the vision encoder and tuned the language encoder to speed up and stabilize the pre-training process, and applied Noah-Wukong [40] and Zero-Corpus [41] as the pre-training datasets. There are two main limitations of existing dataset collection methods. Specifically, a dataset can contain 100 million image pairs collected from the web. The Wukong dataset was compiled from a query list of 200,000 words in order to cover a sufficiently diverse set of visual concepts. Furthermore, the CLIP model transforms this encoded text input into an image embedding, the output of which is also used for a loss function. In addition, this encoded text input is transformed into an image embedding by the CLIP model. This output is also used for a loss function. The specific operation of the model for each input uses a text prompt. It is necessary to retrain a NeRF model, which will require multi-angle 2D photos, and after completing the training can generate a 3D model, and thus, synthesis of the new perspective. The role of CLIP is still to evaluate the accuracy of text-generated images.

5.2. Blockchain Encryption Service

3D Printing has gone through these stages: from concept to CAD file to generating design (if available) to actual 3D Printing. All these steps represent a loophole where 3D Printing could be compromised or even stolen, putting the company’s intellectual property at risk. All projects, from start to finish, can be done in a blockchain, from the communication of the project to the production and transfer of data, and 3D printing and delivery. Everything could be accessible in the chain, and each party would have all the data. Off-chain, cross-chain, off-chain management, and off-chain certification facilities are primarily referred to as “blockchain infrastructure” in this context. They serve as the foundation for communication between blockchains [42]. As shown in Figure 5, firstly, the 3D-printing files are encrypted using a hashing algorithm. Based on the content of the file, the encryption algorithm generates a fingerprint of a unique hash value. The blockchain stores a hash value that verifies the authenticity of the file rather than the file itself. At the second step, we upload the hash value generated after encryption, which is the so-called digital fingerprint, to the blockchain. Finally, if someone wants to print the 3D content, the service will upload the key printing information, such as operator, and location, to the blockchain. Then the blockchain transaction ID is generated, which can further trace the source of the 3D content operation. Trusted printers can communicate with the blockchain by installing so-called Secure Elements [43] onto AM machines.
The characteristics and their relevance to AM are shown in Table 2. The most significant advantage of applying blockchain technology to AM is the development of trust. The designer would no longer need to be concerned about his designs being used illegally and could instead focus on getting better at making models. The individual user could also choose to print the files in a variety of ways, either by the number of times they are printed or by gaining access to the files and downloading them directly, both of which are now available at a relatively lower price and with full disclosure regarding the models’ origin. On-demand production is also environmentally friendly and carbon neutral, and manufacturers could produce locally, close to their customers, thereby reducing storage and transportation costs. In addition to the actual goods, brand manufacturers could offer their customers printable files for customization, accessories, and replacement parts. In short, the blockchain for 3D printing will make it possible to trade 3D printing in the form of “tokens” and promote 3D printing as a technology to a greater extent.

5.3. Cloud Management Service

Theoretically, 3D printing is the ideal method of production for cloud manufacturing. Digital files can be printed anywhere with just a 3D printer and suitable materials, thanks to their ease of transfer and no geographic restrictions. The cloud-based 3D printing product personalization service platform uses a browser/server model that allows users to use a browser to access the platform. The Internet serves as the medium between the browser and the server, enabling the cloud service platform to provide a variety of cloud 3D-Printing services. This cloud platform’s architecture consists primary of three layers [30]: the virtual resources layer, the 3D-printing-resource layer, and the technology support layer. The foundation for the cloud services platform is the primary technical support layer. The cloud administration stage must include the framework as a management of the executives model, which intends to provide consistent and smooth specialized assistance for the cloud-producing administration stage’s activity. The primary function of the virtual resources layer is to abstract and simplify cloud service platform-connected 3D-printing resources: The cloud service platform’s cloud computing technology is used to describe the various physical 3D-printing resources as virtual resources [31], resulting in virtual data resources. A virtual cloud pool is created when virtual data resources are encapsulated and published to the cloud platform’s resource service center module. Users can select the printing resources they require from the cloud. For the personalized service platform, the 3D-printing resource layer provides software, material, and equipment resources. The purpose of the user interface layer is to provide the cloud service platform with user-friendly application interfaces, allowing users to invoke various cloud services freely.

6. Case Study

Recreational equipment manufacturing is selected as the practice object to verify the effectiveness of the product customization framework proposed in this paper. The toy manufacturing industry is an integral part of the traditional manufacturing industry. It has a high demand for labor and a large export volume of products. Surveys show that the 2020 toy and game market is projected to be US $135 billion [44]. 3D Printing allows the creation of physical objects from geometric representations by continuously adding material. It can effectively respond to customers’ needs and help achieve service-oriented manufacturing. In China, the traditional toy market has interactive electronic toys with high-technology content, high-tech intelligent toys, and educational toys. They can foster children’s imagination, creativity, and hands-on skills. These toys are mainstream. The traditional peak season for the toy industry is June to October each year. Thanks to government measures to promote consumption, and urban consumption upgrades, China’s toy and retail scale have maintained steady growth. The main export markets are the United States, the United Kingdom, Japan, and other countries, of which exports to the United States amounted to US $8.57 billion in 2021, an increase in 6.8% over the previous year, accounting for 25.6% of China’s total exports. According to the research study of Made-in-China, China’s Toys Export in November Amounted to US $ 2.44 Billion, up by 21.18%.
In this paper, the pictures used in the production of datasets were obtained by taking video with mobile phones and then drawing frames. The dataset consists of 86 different scenes, which are mainly composed of scenes with four rotations and scans at 90-degree intervals, thereby realizing a 360-degree model. The production of the dataset can be summarized in three steps. First, do feature matching of pictures to obtain the camera pose. Second, convert the matching pose into LLFF format. Finally, upload the required files to the corresponding folder of NeRF and set up the configuration file. LLFF-format data can be the corresponding picture parameters. Camera position and camera parameters can be stored in a simple and effective file to facilitate Python reading. And the NeRF model’s source code has the necessary configuration and modules for direct training on LLFF-format datasets, making it easy for researchers to use. COLMAP is a universal motion structure (SfM) and multi-view stereo (MVS) processing tool with graphical and command-line interfaces. It provides a wide range of functions for the reconstruction of ordered and disordered image collections. We used this tool to estimate the photo position. Through the sparse reconstruction of COLMAP, the position and posture of the photo are restored. The next step is to generate the data format used for NeRF training and select the data in LLFF format. The whole process is end-to-end trainable using the dataset as input. The trained NeRF-based model is used to build an end-to-end module. Users can provide multiple photos at different angles for the same object. For example, take one image at each angle, and provide 3–4 angled pictures. If you need to generate a position that cannot be photographed, such as the bottom, you need to provide further pictures of the model lying on its side. When moving around an object, adjacent photos must overlap by at least 70%. It is recommended to take at least 30 photos to provide more pictures, which is helpful in generating higher-quality 3D models.
As for the text-to-3D module, we use a Pytorch implementation of an original Dreamfields algorithm as baseline. We mainly changes the back end of the original Dreamfields from the original NeRF to our model and replaces the origin CLIP encoder with the Taiyi-CLIP [37] encoder. In addition, to improve the performance of the generated content, we apply R-Drop [45] for regularization in the forward propagation of the CLIP pre-trained model; these methods enhance the expressiveness and generalization ability of our model. We used search engines to crawl about 10,000 image–text pairs of data related to the toy manufacturing industry to form fine-tuned datasets. Each image has up to five descriptions associated with it. The input for the model is a batch of subtitles and a batch of images passed through the CLIP text encoder and image encoder, respectively. The training process uses contrast learning to learn joint embedded representations of images and subtitles. In this embedded space, the images are very close to their respective descriptions, as are similar images and similar descriptions. Conversely, images and descriptions of different images may be pushed further. In order to standardize our dataset and prevent overfitting due to the size of the dataset, we used both image and text enhancement. Image enhancement was done online using the built-in conversion in the Pytorch Torchvision package. The transformations used were random cropping, random resizing and cropping, color jitter, and random horizontal and vertical flipping.
As shown in Table 3 and Table 4, we have performed some experiments to test the print time, quality, and the corresponding actual printing time of the generated models with different inputs, and we can see that the models generated within this framework are reliable in terms of efficiency and quality. Some results are shown in Figure 6.
The resource scheduling interface of the cloud platform adopts the development mode of separating the front and back ends. Springboot, Mybatis, and Springcloud frameworks were applied to develop the back-end of the cloud platforms. The persistent operation of data storage uses MongoDB and MySQL databases. The cloud platform is equipped with a cloud server with 16 G memory and 8 NVIDIA V100 GPUs graphics cards. We deployed the pre-trained large model described above on TF-Serving’s server. TF-Serving will automatically deploy it according to the incoming port and model path. The server where the model is located does not need a Python environment (take Python training as an example). And then the application service directly launches a service call to the server where the model is located. The call can be made through grpc.
The 3D models of the 3D model file library come from the uploads of various designers on the cloud platform. The designers of the 3D printing cloud platform designed 3D models with a certain market value according to their own capabilities and market needs and then uploaded them to the cloud platform. The cloud platform is responsible for optimizing and managing these model files. Users can quickly search for suitable 3D model files. The management of the 3D model file library mainly includes the removing repeat, classification, evaluation, and security management of model files.
The cloud platform’s overall framework is shown in Figure 7. It supports data processing in the 3D printing process, including data format conversion, support design, slice calculation, print path planning, structural analysis, model optimization, etc. These steps depend on each other and affect each other, and together determine the the efficiency of printing and the quality of the finished product. In addition, in order to improve the computational efficiency of the support-generation process in the 3D-printing process, the parallelized slicing method, the model placement pointing optimization and its parallelization method [46], and GPU-based parallelized support generation were successfully applied to this platform.

7. Conclusions

In SM, providing a user-friendly approach to product customization has been the focus of research. As an early attempt to apply NeRF and large pre-trained vision-language models such as CLIP, we propose an efficient product customization framework in the SM paradigm. It provides a new method of collaborative product design, which can efficiently solve the problems existing in current manufacturing systems, such as the accuracy of 3D modeling, security, supervision, anti-counterfeiting, and efficient allocation of resources. However, the key technologies used in this paper, such as neural-volume rendering and multimodal data processing technology, still need to be further studied and improved. The main limitations can be summarized as follows. Firstly, choosing a suitable resolution is very important for printing a high-quality model. A too low resolution will inevitably affect the quality of the finished print. A low resolution results in a non-smooth surface for the finished 3D printing. It is currently difficult to obtain high-resolution geometry or textures for 3D models with the model presented in this paper. To address this issue, a coarse-to-fine optimization approach could be proposed, in which multiple diffusion priors at different resolutions would optimize the 3D representation, which would result in view-consistent geometry and high-resolution details. Secondly, in the current implementation of the 3D model generation algorithm, a NeRF network needs to be retrained for each text prompt, which results in low efficiency of model generation and requires a large amount of GPU resources. In the future, we can consider loading the existing generic mesh model and then iteratively modifying the 3D model by text prompts to generate an ideal 3D model. A priori forward-looking production framework for future integration would be instructive. Accordingly, this work is proved to be a revolutionary break through for solving the core problems of 3D modeling in SM.

Author Contributions

Conceptualization, Y.L. and Z.S.; methodology, Y.L.; software, Y.L. and S.L.; validation, Y.L., Z.S. and G.X.; formal analysis, H.W.; investigation, Y.L. and T.S.T.; writing—original draft preparation, Y.L. and T.S.T.; writing—review and editing, Y.L., Z.S. and B.H.; visualization, Y.L. and Z.S.; funding acquisition, Z.S. and G.X. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Key Research and Development Program of China under Grant 2018YFB1700602; in part by the National Natural Science Foundation of China under Grants 92267103, U1909204 and 61872365; in part by the Scientific Instrument Developing Project of the Chinese Academy of Sciences (CAS) under Grant YZQT014; in part by the Guangdong Basic and Applied Basic Research Foundation under Grant 2021B1515140034; in part by the Foshan Science and Technology Innovation Team Project under Grant 2018IT100142; and in part by the Collaborative Innovation Center of Intelligent Green Manufacturing Technology and Equipment, Shandong, under Grant IGSD-2020-015. The work of Zhen Shen was supported by the CAS Key Technology Talent Program.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ding, K.; Jiang, P. Incorporating social sensors, cyber-physical system nodes, and smart products for personalized production in a social manufacturing environment. Proc. Inst. Mech. Eng. Part B J. Eng. Manuf. 2018, 232, 2323–2338. [Google Scholar] [CrossRef]
  2. Zheng, T.; Ardolino, M.; Bacchetti, A.; Perona, M. The applications of Industry 4.0 technologies in manufacturing context: A systematic literature review. Int. J. Prod. Res. 2021, 59, 1922–1954. [Google Scholar] [CrossRef]
  3. Wang, F.Y. The emergence of intelligent enterprises: From CPS to CPSS. IEEE Intell. Syst. 2010, 25, 85–88. [Google Scholar] [CrossRef]
  4. Wang, F.Y. From social computing to social manufacturing: The coming industrial revolution and new frontier in cyber-physical-social space. Bull. Chin. Acad. Sci. 2012, 6, 658–669. [Google Scholar]
  5. Wang, F.Y. Social computing: Concepts, contents, and methods. Int. J. Intell. Control Syst. 2004, 9, 91–96. [Google Scholar]
  6. Xiong, G.; Wang, F.Y.; Nyberg, T.R.; Shang, X.; Zhou, M.; Shen, Z.; Li, S.; Guo, C. From mind to products: Towards social manufacturing and service. IEEE/CAA J. Autom. Sin. 2017, 5, 47–57. [Google Scholar] [CrossRef]
  7. Shang, X.; Wang, F.Y.; Xiong, G.; Nyberg, T.R.; Yuan, Y.; Liu, S.; Guo, C.; Bao, S. Social manufacturing for high-end apparel customization. IEEE/CAA J. Autom. Sin. 2018, 5, 489–500. [Google Scholar] [CrossRef]
  8. Shang, X.; Shen, Z.; Xiong, G.; Wang, F.Y.; Liu, S.; Nyberg, T.R.; Wu, H.; Guo, C. Moving from mass customization to social manufacturing: A footwear industry case study. Int. J. Comput. Integr. Manuf. 2019, 32, 194–205. [Google Scholar] [CrossRef]
  9. Shang, X.; Chen, X.; Niu, L.; Xiong, G.; Shen, Z.; Dong, X.; Shen, Z.; Liu, C.; Xi, B. Blockchain-based social manufacturing for customization production. IFAC-PapersOnLine 2020, 53, 53–58. [Google Scholar] [CrossRef]
  10. Mohajeri, B.; Nyberg, T.; Karjalainen, J.; Tukiainen, T.; Nelson, M.; Shang, X.; Xiong, G. The impact of social manufacturing on the value chain model in the apparel industry. In Proceedings of the 2014 IEEE International Conference on Service Operations and Logistics, and Informatics, Qingdao, China, 8–10 October 2014; pp. 378–381. [Google Scholar]
  11. Mohajeri, B.; Nyberg, T.; Karjalainen, J.; Nelson, M.; Xiong, G. Contributions of social manufacturing to sustainable apparel industry. In Proceedings of the 2016 IEEE International Conference on Service Operations and Logistics, and Informatics (SOLI), Beijing, China, 10–12 July 2016; pp. 24–28. [Google Scholar]
  12. Mohajeri, B.; Kauranen, I.; Nyberg, T.; Ilen, E.; Nelson, M.; Xiong, G. Improving sustainability in the value chain of the apparel industry empowered with social manufacturing. In Proceedings of the 2020 15th IEEE Conference on Industrial Electronics and Applications (ICIEA), Kristiansand, Norway, 9–13 November 2020; pp. 235–240. [Google Scholar]
  13. Jiang, P.; Leng, J.; Ding, K. Social manufacturing: A survey of the state-of-the-art and future challenges. In Proceedings of the 2016 IEEE International Conference on Service Operations and Logistics, and Informatics (SOLI), Beijing, China, 10–12 July 2016; pp. 12–17. [Google Scholar]
  14. Ding, K.; Jiang, P.Y.; Zhang, X. A framework for implementing social manufacturing system based on customized community space configuration and organization. In Advanced Materials Research; Trans Tech Publications Ltd.: Bäch, Switzerland, 2013; Volume 712, pp. 3191–3194. [Google Scholar]
  15. Jiang, P.; Ding, K.; Leng, J. Towards a cyber-physical-social-connected and service-oriented manufacturing paradigm: Social Manufacturing. Manuf. Lett. 2016, 7, 15–21. [Google Scholar] [CrossRef]
  16. Leng, J.; Jiang, P.; Xu, K.; Liu, Q.; Zhao, J.L.; Bian, Y.; Shi, R. Makerchain: A blockchain with chemical signature for self-organizing process in social manufacturing. J. Clean. Prod. 2019, 234, 767–778. [Google Scholar] [CrossRef]
  17. Xiong, G.; Helo, P.; Ekstrom, S.; Tamir, T.S. A Case Study in Social Manufacturing: From social manufacturing to social value chain. Machines 2022, 10, 978. [Google Scholar] [CrossRef]
  18. Cao, W.; Jiang, P.; Jiang, K. Demand-based manufacturing service capability estimation of a manufacturing system in a social manufacturing environment. Proc. Inst. Mech. Eng. Part B J. Eng. Manuf. 2017, 231, 1275–1297. [Google Scholar] [CrossRef]
  19. Xiong, G.; Chen, Y.; Shang, X.; Liu, X.; Nyberg, T.R. AHP fuzzy comprehensive method of supplier evaluation in social manufacturing mode. In Proceedings of the the 11th World Congress on Intelligent Control and Automation, Shenyang, China, 29 June–4 July 2014; pp. 3594–3599. [Google Scholar]
  20. Yin, D.; Ming, X.; Zhang, X. Understanding data-driven cyber-physical-social system (D-CPSS) using a 7C framework in social manufacturing context. Sensors 2020, 20, 5319. [Google Scholar] [CrossRef]
  21. Xiong, G.; Tamir, T.S.; Shen, Z.; Shang, X.; Wu, H.; Wang, F.Y. A Survey on Social Manufacturing: A Paradigm Shift for Smart Prosumers. IEEE Trans. Comput. Soc. Syst. 2022, 1–19. [Google Scholar] [CrossRef]
  22. Brière-Côté, A.; Rivest, L.; Maranzana, R. Comparing 3D CAD models: Uses, methods, tools and perspectives. Comput.-Aided Des. Appl. 2012, 9, 771–794. [Google Scholar] [CrossRef] [Green Version]
  23. Roberts, L.G. Machine Perception of Three-Dimensional Solids. Ph.D. Thesis, Massachusetts Institute of Technology, Cambridge, MA, USA, 1963. [Google Scholar]
  24. Yao, Y.; Luo, Z.; Li, S.; Fang, T.; Quan, L. Mvsnet: Depth inference for unstructured multi-view stereo. In Proceedings of the the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 767–783. [Google Scholar]
  25. Mildenhall, B.; Srinivasan, P.P.; Tancik, M.; Barron, J.T.; Ramamoorthi, R.; Ng, R. Nerf: Representing scenes as neural radiance fields for view synthesis. Commun. ACM 2021, 65, 99–106. [Google Scholar] [CrossRef]
  26. Mai, J.; Zhang, L.; Tao, F.; Ren, L. Customized production based on distributed 3D printing services in cloud manufacturing. Int. J. Adv. Manuf. Technol. 2016, 84, 71–83. [Google Scholar] [CrossRef]
  27. Adamson, G.; Wang, L.; Holm, M. The state of the art of cloud manufacturing and future trends. In Proceedings of the International Manufacturing Science and Engineering Conference, Madison, WI, USA, 10–14 June 2013; American Society of Mechanical Engineers: New York, NY, USA, 2013; Volume 55461, p. V002T02A004. [Google Scholar]
  28. Xu, X. From cloud computing to cloud manufacturing. Robot. Comput.-Integr. Manuf. 2012, 28, 75–86. [Google Scholar] [CrossRef]
  29. Tao, F.; Zhang, L.; Venkatesh, V.; Luo, Y.; Cheng, Y. Cloud manufacturing: A computing and service-oriented manufacturing model. Proc. Inst. Mech. Eng. Part B J. Eng. Manuf. 2011, 225, 1969–1976. [Google Scholar] [CrossRef]
  30. Guo, L.; Qiu, J. Combination of cloud manufacturing and 3D printing: Research progress and prospect. Int. J. Adv. Manuf. Technol. 2018, 96, 1929–1942. [Google Scholar] [CrossRef]
  31. Cui, J.; Ren, L.; Mai, J.; Zheng, P.; Zhang, L. 3D printing in the context of cloud manufacturing. Robot. Comput.-Integr. Manuf. 2022, 74, 102256. [Google Scholar] [CrossRef]
  32. Yang, C.; Wang, Y.; Tang, R.; Lan, S.; Wang, L.; Shen, W.; Huang, G.Q. Cloud-edge-device Collaboration Mechanisms of Cloud Manufacturing for Customized and Personalized Products. In Proceedings of the 2022 IEEE 25th International Conference on Computer Supported Cooperative Work in Design (CSCWD), Hangzhou, China, 4–6 May 2022; pp. 1517–1522. [Google Scholar]
  33. Tamir, T.S.; Xiong, G.; Dong, X.; Fang, Q.; Liu, S.; Lodhi, E.; Shen, Z.; Wang, F.Y. Design and optimization of a control framework for robot assisted additive manufacturing Based on the Stewart Platform. Int. J. Control Autom. Syst. 2022, 20, 968–982. [Google Scholar] [CrossRef]
  34. Müller, T.; Evans, A.; Schied, C.; Keller, A. Instant neural graphics primitives with a multiresolution hash encoding. arXiv 2022, arXiv:2201.05989. [Google Scholar] [CrossRef]
  35. Radford, A.; Kim, J.W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J.; et al. Learning transferable visual models from natural language supervision. In Proceedings of the International Conference on Machine Learning(ICML), Online, 18–24 July 2021; pp. 8748–8763. [Google Scholar]
  36. Jain, A.; Mildenhall, B.; Barron, J.T.; Abbeel, P.; Poole, B. Zero-shot text-guided object generation with dream fields. In Proceedings of the the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 21–24 June 2022; pp. 867–876. [Google Scholar]
  37. Wang, J.; Zhang, Y.; Zhang, L.; Yang, P.; Gao, X.; Wu, Z.; Dong, X.; He, J.; Zhuo, J.; Yang, Q.; et al. Fengshenbang 1.0: Being the foundation of chinese cognitive intelligence. arXiv 2022, arXiv:2209.02970. [Google Scholar]
  38. Cui, Y.; Che, W.; Liu, T.; Qin, B.; Yang, Z. Pre-training with whole word masking for chinese bert. IEEE/ACM Trans. Audio Speech Lang. Process. 2021, 29, 3504–3514. [Google Scholar] [CrossRef]
  39. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  40. Gu, J.; Meng, X.; Lu, G.; Hou, L.; Niu, M.; Xu, H.; Liang, X.; Zhang, W.; Jiang, X.; Xu, C. Wukong: 100 Million large-scale Chinese cross-modal pre-training dataset and a foundation framework. arXiv 2022, arXiv:2202.06767. [Google Scholar]
  41. Xie, C.; Cai, H.; Song, J.; Li, J.; Kong, F.; Wu, X.; Morimitsu, H.; Yao, L.; Wang, D.; Leng, D.; et al. Zero and R2D2: A large-scale Chinese cross-modal benchmark and a vision-Language framework. arXiv 2022, arXiv:2205.03860. [Google Scholar]
  42. Ouyang, L.; Yuan, Y.; Wang, F.Y. A blockchain-based framework for collaborative production in distributed and social manufacturing. In Proceedings of the 2019 IEEE International Conference on Service Operations and Logistics, and Informatics (SOLI), Zhengzhou, China, 6–8 November 2019; pp. 76–81. [Google Scholar]
  43. Holland, M.; Nigischer, C.; Stjepandić, J. Copyright protection in additive manufacturing with blockchain approach. In Transdisciplinary Engineering: A Paradigm Shift; IOS Press: Amsterdam, The Netherlands, 2017; pp. 914–921. [Google Scholar]
  44. Petersen, E.E.; Kidd, R.W.; Pearce, J.M. Impact of DIY home manufacturing with 3D printing on the toy and game market. Technologies 2017, 5, 45. [Google Scholar] [CrossRef] [Green Version]
  45. Liang, X.; Wu, L.; Li, J.; Wang, Y.; Meng, Q.; Qin, T.; Chen, W.; Zhang, M.; Liu, T.-Y. R-drop: Regularized dropout for neural networks. Adv. Neural Inf. Process. Syst. 2021, 34, 10890–10905. [Google Scholar]
  46. Li, Z.; Xiong, G.; Zhang, X.; Shen, Z.; Luo, C.; Shang, X.; Dong, X.; Bian, G.B.; Wang, X.; Wang, F.Y. A GPU based parallel genetic algorithm for the orientation optimization problem in 3D printing. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 2786–2792. [Google Scholar]
Figure 1. An overview of NeRF model presented by Ben Mildenhall et al. [25]. We create images using the sampling of 5D coordinates (location and viewing direction) along camera rays (a), feeding those locations into an MLP to create a color and volume density (b), and then utilizing volume rendering algorithms to merge these values into an image (c). Because of the differentiability of this rendering function, we can improve the scene representation by decreasing the difference between the synthetic and actual ground truth observed images (d).
Figure 1. An overview of NeRF model presented by Ben Mildenhall et al. [25]. We create images using the sampling of 5D coordinates (location and viewing direction) along camera rays (a), feeding those locations into an MLP to create a color and volume density (b), and then utilizing volume rendering algorithms to merge these values into an image (c). Because of the differentiability of this rendering function, we can improve the scene representation by decreasing the difference between the synthetic and actual ground truth observed images (d).
Machines 11 00170 g001
Figure 2. The Efficient Product Customization Framework.
Figure 2. The Efficient Product Customization Framework.
Machines 11 00170 g002
Figure 3. The pipeline of 3D content generation service.
Figure 3. The pipeline of 3D content generation service.
Machines 11 00170 g003
Figure 4. The schematic diagram of the principle of CLIP model.
Figure 4. The schematic diagram of the principle of CLIP model.
Machines 11 00170 g004
Figure 5. Process of blockchain encryption of 3D contents.
Figure 5. Process of blockchain encryption of 3D contents.
Machines 11 00170 g005
Figure 6. 3D content generation examples.
Figure 6. 3D content generation examples.
Machines 11 00170 g006
Figure 7. The product customization framework built on a cloud platform.
Figure 7. The product customization framework built on a cloud platform.
Machines 11 00170 g007
Table 1. Advantages and disadvantages of the mainstream 3D modeling approach.
Table 1. Advantages and disadvantages of the mainstream 3D modeling approach.
MethodsAdvantagesDisadvantages
Forward Modeling (CAD)Ideal for showing details of objectsComplicated and challenging to parameterize accurately
MVS 3D ReconstructionBetter performance when geometric information is abundant and accurateMissing or incorrect geometry reduces the quality of the view.
NeRF-based 3D ReconstructionHigher quality modelsLong training time and poor generalization.
Table 2. Blockchain characteristics and their relevance to AM.
Table 2. Blockchain characteristics and their relevance to AM.
Blockchain CharacteristicRelevance to AM
Distributed data maintained within and among stakeholdersHelps in the management of activity in distributed supply chain expected to be found with AM
Near real time-settlement and exchange are nearly instantChanges to a design are made instantly, which facilitates efficient AM processes
Trustless environment—cryptographic validation of transactionsDesigned to protect against risks of unauthorized data access
Irreversibility the transaction history is append-onlyHelps with cyber risk and lP protection as it is intended to provide an indelible and traceable record of changes
Table 3. Some examples of 3D content generation based on Surrounding photos.
Table 3. Some examples of 3D content generation based on Surrounding photos.
Product Name (Numbers of Photos)Generate TimePrint Time
“Gundam (74)”17 min 58 s2 h 38 min 30 s
“Plane (102)”13 min 12 s1 h 5 min 22 s
“Princess (194)”40 min 23 s3 h 43 min 28 s
“Doll Cow (140)”21 min 25 s4 h 48 min 52 s
Table 4. Some examples of 3D content generation based on text prompt.
Table 4. Some examples of 3D content generation based on text prompt.
Text PromptGenerate TimePrint Time
“An unlighted candle.”45 min 37 s1 h 55 min 26 s
“A science fiction style gear”41 min 35 s3 h 48 min 4 s
“A chair”28 min 47 s3 h 58 min 11 s
“A plush toy of a corgi”33 min 12 s3 h 50 min 37 s
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Y.; Wu, H.; Tamir, T.S.; Shen, Z.; Liu, S.; Hu, B.; Xiong, G. An Efficient Product-Customization Framework Based on Multimodal Data under the Social Manufacturing Paradigm. Machines 2023, 11, 170. https://doi.org/10.3390/machines11020170

AMA Style

Li Y, Wu H, Tamir TS, Shen Z, Liu S, Hu B, Xiong G. An Efficient Product-Customization Framework Based on Multimodal Data under the Social Manufacturing Paradigm. Machines. 2023; 11(2):170. https://doi.org/10.3390/machines11020170

Chicago/Turabian Style

Li, Yanpeng, Huaiyu Wu, Tariku Sinshaw Tamir, Zhen Shen, Sheng Liu, Bin Hu, and Gang Xiong. 2023. "An Efficient Product-Customization Framework Based on Multimodal Data under the Social Manufacturing Paradigm" Machines 11, no. 2: 170. https://doi.org/10.3390/machines11020170

APA Style

Li, Y., Wu, H., Tamir, T. S., Shen, Z., Liu, S., Hu, B., & Xiong, G. (2023). An Efficient Product-Customization Framework Based on Multimodal Data under the Social Manufacturing Paradigm. Machines, 11(2), 170. https://doi.org/10.3390/machines11020170

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop