Next Article in Journal
Practical Use of Starlink Downlink Tones for Positioning
Previous Article in Journal
Attitude-Orbit Coupled Control of Gravitational Wave Detection Spacecraft with Communication Delays
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Preprocessing for Keypoint-Based Sign Language Translation without Glosses

Department of Computer Science and Engineering, Incheon National University (INU), Incheon 22012, Republic of Korea
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(6), 3231; https://doi.org/10.3390/s23063231
Submission received: 22 February 2023 / Revised: 12 March 2023 / Accepted: 15 March 2023 / Published: 17 March 2023
(This article belongs to the Section Intelligent Sensors)

Abstract

:
While machine translation for spoken language has advanced significantly, research on sign language translation (SLT) for deaf individuals remains limited. Obtaining annotations, such as gloss, can be expensive and time-consuming. To address these challenges, we propose a new sign language video-processing method for SLT without gloss annotations. Our approach leverages the signer’s skeleton points to identify their movements and help build a robust model resilient to background noise. We also introduce a keypoint normalization process that preserves the signer’s movements while accounting for variations in body length. Furthermore, we propose a stochastic frame selection technique to prioritize frames to minimize video information loss. Based on the attention-based model, our approach demonstrates effectiveness through quantitative experiments on various metrics using German and Korean sign language datasets without glosses.

1. Introduction

Effective communication is essential for full participation in society, yet deaf individuals often face challenges in accessing information and communicating with hearing unimpaired individuals [1]. Sign language is a visual language that provides the primary means of communication for deaf individuals. It uses a combination of hand movements, facial expressions, mouth movements, and upper body movements to convey information. However, sign language is often not well understood by those unfamiliar with it, making it difficult for deaf individuals to communicate with the general population. In order to address this issue, research on sign language translation (SLT) using computers has been conducted for some time to facilitate communication between deaf and non-deaf people via translating sign language videos into spoken language [2,3].
Most existing deep-learning-based SLT methods rely on gloss annotations, a minimum vocabulary set of expressions that follow the sign language sequence and grammar, to translate the sign language into spoken language. However, collecting gloss annotations can be costly and time-consuming, and there is a need for SLT methods that do not require them. The state-of-the-art study [4] attempted to perform translation without glosses, but came across two major limitations. Firstly, the translation was based on the features of the image, which resulted in it being influenced by the background rather than focusing solely on the signer. Secondly, dimensionality reduction was applied to the features through embedding to fix the input size of the model. However, this may result in the loss of important information in sign language videos.
To address the first issue, we propose a sign language translation method based on the body movement of the signer. Sign language represents a visual language essential for communication among the deaf population, and it uses various complementary channels to convey information [5]. These channels include hand movements and nonverbal elements such as the signer’s facial expressions, mouth, and upper body movements, which are crucial for effective communication [6]. Therefore, the meaning of sign language can vary depending on the position and space of the signer’s hand. Taking these factors into consideration, we focus on the signer’s movement by accurately capturing their body’s movement using keypoints, namely, specific points on the body that are used to define the movement of the signer. These keypoints are defined as the joint points of these signers.
To process the frames for SLT based on the signer’s skeleton, we must address the issue of keypoint vectors that vary significantly depending on the angle and position of the signer in the frame. To address this issue, we propose a simple but effective method. First, we propose a normalization method that uses the distance between keypoints and applies different normalization methods according to the part of the body, which we call “Customized Normalization”. This normalization method is robust to variations in the signer’s body and helps improve the performance of the translation system.
To solve the second issue of previous studies [4], we propose selecting and augmenting frames based on probabilities. Our method can reduce the loss of image information due to embedding by prioritizing frames in sign language images. Previous keypoint-based studies [7,8,9,10] have also employed various methods to fix the length of the input keypoint vector. However, these studies either use frame sampling to lose information or use only augmentation, leading to increased memory usage and slower processing time. Therefore, we propose the“Stochastic Augmentation and Skip Sampling (SASS)” method that adjusts to the length of a dynamic video frame by simultaneously using both augmentation and sampling techniques based on the length of the video.
In this paper, we apply the Sign2Text method to the dataset without gloss and propose a frame-processing method that can perform well even in cases where the video resolution differs. For sign language translation, we use the Gated Recurrent Unit (GRU) [11]-based Sequence-to-Sequence (Seq2Seq) model [12] with the Attention of Bahdanau et al. [13]. An overview of our framework is illustrated in Figure 1, and the contributions of this paper are as follows.
We propose a new normalization method most suitable for sign language translation in keypoint processing.
We propose a stochastic frame selection method that prioritizes frames for videos of different lengths to consider the characteristics of a sign language video.
Our method is versatile because it can be applied to datasets without gloss.
The remainder of this paper is structured as follows: Section 2 reviews existing research on SLT and video processing methods. Section 3 presents our proposed video processing method and SLT approach. Section 4 presents the results of experiments evaluating the effectiveness of our method. Section 5 presents the results of the ablation study. Finally, Section 6 provides concluding remarks and suggestions for future work.

2. Related Works

2.1. Sign Language Translation

As mentioned earlier, research on sign language translation has been steadily progressing, but there are several reasons for the slow development of sign language translation systems. First, there needs to be more data for sign language translation. Since the sign language differs depending on the country and region, the data to translate a specific language differs. Therefore, it is necessary to collect sign language datasets directly by country, but collecting many datasets takes much work. In previous studies that only dealt with translating between sign language and text, the average number of words in the dataset was very small, about 3000, even though no image information was used [14,15,16]. Fortunately, researchers began building many sign language datasets as algorithms evolved. In particular, with the development of algorithms for weakly annotation data [17,18] and the development of algorithms for pose estimation [19,20], researchers began to feel the need for sign language datasets.
As a result of these developments, German sign language datasets called RWTH-PHOENIX-Weather 2012 [21], RWTH-PHOENIX-Weather 2014 [22], RWTH-PHOENIX-Weather 2014-T [4], an American Sign Language (ASL) dataset ASL-PC12 [23], and Korean sign language dataset KETI [7] were also created.
The establishment of various datasets has led to a surge in deep learning-based SLT research. The PHOENIX dataset, which includes glosses, has been particularly popular and generated much interest and research. Most of these studies extract sign features from sign language videos using Convolution Neural Networks (CNNs) [24]. For example, Yin et al. [25] combined the STMC [26] module, which extracts the signer’s features and separates body parts, with a Transformer [27] based on CNN. Camgoz et al. [4,28] also used a CNN-based model to extract the signer’s features from videos.
However, the signer’s features can also be extracted from sign language videos based on their skeleton movements, not just using CNN. Gan et al. [10] translated sign language from the PHOENIX dataset using GCN [29], based on the signer’s keypoints. Additionally, research by Ko et al. [7] and Kim et al. [8] also translated Korean sign language based on the signer’s keypoints.

2.2. Video Processing

Video processing is mainly applied in action recognition tasks and video recognition tasks, with video processing having a significant impact on performance [30,31,32]. Gowda et al. [33] proposed a score-based frame selection method and achieved state-of-the-art results for the UCF-101 dataset [34]. Karpathy et al. [35] performed video classification on a large-scale dataset using a video-processing method that treats all videos as fixed clips. Therefore, video processing is a crucial method for performance improvement.
Video processing mainly uses an approach that changes the number of frames, via augmenting the number of frames by transforming the frame image [36,37] or augmenting the manipulation of the order of the frames [38]. Recently, some studies have proposed a frame augmentation method using Generative Adversarial Networks (GAN) [39] or genetic algorithms for video processing using deep learning [40]. In addition to augmentation, a frame sampling method, frame skipping [41], has also been proposed.
The frame-processing method was employed in both the SLT and the action recognition. Park et al. [9] augmented the data in three ways: camera angle transformation, finger length transformation, and arbitrary keypoint removal. Ko et al. [7] used a new skip sampling method to augment the number of data by adding randomness.

3. Proposed Method

In this section, we introduce the overall architecture of the proposed method. Our architecture is decomposed into four parts: Extraction of keypoint, keypoint normalization, Stochastic Augmentation and Skip Sampling (SASS) and sign language translation model. The overall flow of our proposed architecture is described in Figure 2.
Camgoz et al. [4] underwent the embedding process after extracting the feature of the sign language video frame through CNN. It aims to reduce the dimension by embedding the feature point of the large amount extracted from the learning image. However, all the features extracted can include backgrounds, which can affect learning by becoming noise. Therefore, in this paper, instead of extracting features through CNN, we propose a method of extracting the keypoint of the signer and using it as a feature value. Using keypoint, we can construct a robust model against the background by only looking at the movement. Initially, the location of keypoints is taken into consideration, followed by the proposal of Customized Normalization that emphasizes hand movement, an essential aspect of sign language. In many languages, verbs and nouns, the core of a sentence, are placed in the middle. Furthermore, there are cases in which hands do not appear at the beginning and end of the video. Therefore, we define the ’key frame’ as the frame in the middle part of the video. We propose a stochastic augmentation method that enables the model to better learn essential parts of sign language videos by assigning stochastic weights to ’key frames’ of videos. Additionally, we propose Stochastic Augmentation and Skip Sampling (SASS), which combines sampling and augmentation to apply to dynamic video frame length. Finally, we perform sign language translation by applying the method of Bahdanau et al. [13] of the NMT task.

3.1. Extraction of Keypoint

To minimize the impact of the surrounding background on the video and focus solely on the signer, we extracted the signer’s keypoint using a skeleton point extraction method. Our approach involved using Alphapose, a framework developed by Fang et al. [20] that employs a top-down detection technique for signers and then extracts keypoints from cropped images. We used a pretrained model from the Halpe dataset [20], which has 136 keypoints. By removing the keypoints of the lower body and face keypoints to construct the keypoint vector V = ( v 1 , , v i , , v 55 ) . The location of these keypoints are shown in Figure 3.

3.2. Customized Normalization

We propose customized normalization according to body parts considering skeleton positions. Since the body length varies depending on the person, a normalization method considering this is needed. Although Kim et al. [8] mentioned the robust keypoint normalization method, it cannot be said to be a normalization method considering all positions because the reference point was set in some parts, not all parts. Therefore, we propose a normalization method that can further emphasize the location of each skeleton and also the movement of the hand.
We aim to normalize the keypoint vector, V. The V is divided into x and y coordinates, V x and V y , and normalized, respectively. First, we designated reference points for each body part according to the face, upper body, left arm, and right arm. The reference point is defined as r = ( r x , r y ) . The corresponding keypoint numbers according to each body part are shown in Table 1, and the names corresponding to each number can be referred to in Figure 3. Furthermore, the representative value for the entire body of the signer is defined as center point, ( c x , c y ) , and follows:
c x = 1 55 i = 1 55 v i x , c y = 1 55 i = 1 55 v i y w h e r e , v x V x , v y V y
After obtaining the Euclidean distance, d p o i n t , between the center coordinate, c, and the reference point, r, of each part, it is normalized based on it. It follows:
S p o i n t = { N o s e , N e c k , L E l b o w , R E l b o w } d p o i n t = c x r x 2 + c y r y 2 w h e r e , r x , r y S p o i n t ( V ¯ p o i n t x , V ¯ p o i n t y ) = V x c x d p o i n t , V y c y d p o i n t
Therefore, the keypoint for the body excluding both hands follows, ( V ¯ p o i n t x , V ¯ p o i n t y ) .
Next, we apply min-max scaling to the keypoints of both hands in order to make them equal in scale and compare their movements equally. Min-max scaling adjusts the range to [0,1] by dividing the difference between the maximum and minimum values of the coordinates by subtracting the minimum values from the corresponding coordinates. Here, to avoid encountering the vanishing gradient problem caused by a minimum value of 0, the range was adjusted to [ 0.5 , 0.5 ] by subtracting 0.5 . Therefore, we define the normalized hand keypoint vectors, V ¯ h a n d x and V ¯ h a n d y , by min-max scaling the keypoint and subtracting 0.5 from it, which follows:
( V ¯ h a n d x , V ¯ h a n d y ) = ( V x V m i n x V m a x x V m i n x 0.5 , V y V m i n y V m a x y V m i n y 0.5 )
where, V m a x and V m i n are the maximum and minimum values of the hand keypoint, respectively. Consequently, each and point subjected to our normalization method follows vectors, V x * = [ V ¯ p o i n t x ; V ¯ h a n d x ] and V y * = [ V ¯ p o i n t y ; V ¯ h a n d y ] , and the final keypoint vector is a concatenation of them, V * = [ V x * ; V y * ] .

3.3. Stochastic Augmentation and Skip Sampling (SASS)

To train a deep learning model, it is necessary to fix the input size. Previous studies have used embedding to reduce the dimension and fix the size of the input. However, this can result in the loss of important information from the video. Therefore, we propose a method that preserves the information in the sign language video without using embedding. We achieve this by adjusting the input value length through augmentation or sampling the frames. We propose the SASS (Stochastic Augmentation and Skip Sampling) method, which is a general technique that can be applied to various datasets and emphasizes the key frames of sign language videos without losing them. This is achieved through a combination of an augmentation method that considers frame priority and a sampling method that reduces frame loss. We define the j-th training video as x j = { x t } t = 0 T and L as the number of training videos, where j [ 1 , L ] . We select N frames from between the frames of training video x j , which follows:
N = 1 L j = 1 L x j
If the number of video frames x j is bigger than N, the sampling is used, and if the number of the video frame is less than N, the number of frames in each video is adjusted to N using augmentation. Therefore, input value, X, follows X R L × N × | V * | . Our method can prevent the loss of video information, which is a disadvantage of frame sampling, and the disadvantage of frame augmentation, which can prevent a lot of memory consumption. The overall flow of our SASS method follows Figure 4.

3.3.1. Stochastic Augmentation

Our proposed method does not follow uniform distribution when selecting the frame but rather reconstructs the key frame into a probability distribution that can be preferentially extracted with a high probability of being selected.
The probability distribution that increases the priority for the key frame is constructed from a combination of binomial distributions. The frame selection probability set, F p , has a size equal to the length of the video, T, and follows:
F p = { f ( 1 ; T , p ) , f ( 2 ; T , p ) , , f ( T ; T , p ) }
where,
f ( k ; T , p ) = T 1 k p k ( 1 p ) T k w h e r e , k [ 1 , T ]
We prioritize frames with this set of probabilities. Furthermore, we construct a set, P r n , for probability with a binomial distribution, p, rather than fixing the probability to a single value. The set P r n follows:
P r 0 = { 1 2 } P r n = P r n 1 { 1 n + 2 , n + 1 n + 2 } p P r n
In other words, the probability, p, follows the probability set, P r n , and F p is reconstructed accordingly. The reconstructed final frame selection probability set, F p * is as follows.
F p * = 1 l p p Pr n F p
where, l p is length of P r n . However, F p * cannot augment the middle part of the video, which is the keyframe we defined. Therefore, we rearrange F p * around the median to increase the selection probability for the middle part of the video. That is, if the frame order, k, is smaller than T / 2 , it is sorted in ascending order, and if it is larger, it is sorted in descending order. Therefore, a final set of frame selection probabilities is produced, and based on this, priority probabilities for frame augmentation order are obtained. Figure 5 shows the probability distribution and kurtosis of F p * according to l p . As l p increases, kurtosis increases.

3.3.2. Skip Sampling

We chose a sampling method to avoid missing key frames when selecting a frame. Ko et al. proposed a random sampling method that does not lose the keyframe. Therefore, this paper also conducts sampling according to this method. When a fixed size is N and the number of frames in the current video is T, the average difference z of frames follows:
c = T N 1
Through z, the baseline sequence S n of the video frame follows:
S N = s + ( N 1 ) c w h e r e s = T c ( N 1 ) 2 { S N } N N N
A random sequence R in the [1, z] range is added to the baseline sequence. At this time, the baseline sequence does not exceed the T. With this skip sampling method, we do not lose key frames, and the order of frames is maintained.

3.4. Sign Language Translation

We used a GRU-based model with a sequence-to-sequence (Seq2Seq) structure with an encoder and decoder in the translation step and improved translation performance by adding attention. This model is widely used in NMT and is suitable for translation tasks with variable-length inputs and outputs. We used Bahdanau attention [32].
Unlike encoding the source sentence in the NMT task, we encode a video’s keypoint vector, x = ( V 1 * , , V t * , , V T * ) , and predict the target sentence, y = ( y 1 , , y i , , y T y ) . First, the encoder calculates the hidden state and generates a context vector, which follows:
h t = f ( v t * , h t 1 ) c = q ( h 1 , , h T )
where, h t is the hidden state calculated in time step t and c is the context vector generated through the encoder a nonlinear function. In the decoder, the joint probability is calculated in the process of predicting words, and this follows.
p ( y ) = i = 1 T y p ( y i | { y 1 , y 2 , , y i 1 } , c )
The conditional probability at each time point used in the above equation is as follows.
p ( y i | y 1 , y 2 , , y i 1 , x ) = softmax ( g ( s i ) )
where, s i is the hidden state calculated from the time i of the decoder and follows:
s i = f ( y i 1 , s i 1 , c )
The context vector c i is calculated using the weight for h i and follows:
c i = t = 1 T = a t i h t
where,
a t i = exp ( score ( s i 1 , h t ) ) t = 1 T exp ( score ( s i 1 , h t )
where, the score is an alignment function that calculates the match score between the hidden state of the encoder and the decoder.

4. Experiments

To prove the effectiveness of our proposed method, we conduct several experiments on two datasets: KETI [7], RWTH-PHOENIX-Weather 2014 T [21]. KETI dataset is a korean sign language video consisting of 105 sentences and 419 words and has a full high-definition (HD) video. There were 10 signers, and they filmed the video from two angles. However, the KETI dataset is not provided with a test set, so we randomly split the dataset at a ratio of 8:1:1. The split of training, dev, and test videos are 6043, 800, and 801, respectively. We divided these videos into frames at 30 fps. The KETI dataset did not provide glosses.
The RWTH-PHOENIX-Weather-2014-T dataset is a data set extended from the existing RWTH-PHOENIX-Weather 2014 [23] with the public german sign language. The split of training, dev, and test videos are 7096, 519, and 642, respectively. It has no overlap with the RWTH-PHOENIX-Weather 2014.
The average number of frames of training videos in these two datasets is 153 and 116 frames, respectively. Therefore, the input dataset, X, follows X kor R 6403 × 153 × 110 and X ger R 6408 × 116 × 110 ,respectively.
Our experiment was conducted in NVIDIA RTX A6000 and AMD EPYC 7302 16-Core environment for CPU. Our model was constructed using Pytorch [42], Adam Optimizer [43], and Cross-Entropy loss. The learning rate was set to 0.001 and epochs 100, and the dropout was set to 0.5 for preventing overfitting. Finally, the dimension of hidden states was 512.
We tokenize differently according to the characteristics of Korean and German. In Korean, the KoNLPy [44] library’s Mecab part-of-speech (POS) tagger was used, and in German, the tokenization was performed through the nltk library [45]. Finally, our evaluation metric evaluates our model in three ways: BLUE-4 [46], ROUGE-L [47], and METEOR [48] according to the metric of NMT.
Our experiments focused on four aspects. The first aspect involved conducting comparative experiments to evaluate the performance of different normalization methods. The second aspect entailed testing the effectiveness of SASS by experimenting with video sampling and augmentation. The third aspect explored the impact of changes to the length of set P r n (i.e., l p ), the representative value N, and the sequence of input values. Lastly, we compared our model’s performance to previous studies to confirm its effectiveness.

4.1. Effect of Normalization

This experiment proves that our method is robust compared to various normalization methods. The results of this are shown in Table 2. We fixed and experimented with all factors except the normalization method. We compared our method with “Standard Normalization”, “Min-Max Normalization” and “Robust Normalization”. First, standard normalization uses standard deviation, σ , and mean values, μ , of points, x, and follows (17). Second, min-max normalization uses the minimum value, x min , and maximum value, x max , of points and follows (18). Finally, robust normalization uses median, x 2 / 4 , and interquartile range (IQR), x 3 / 4 x 1 / 4 , and follows (19).
X = x μ σ
X = x x max x max x min
X = x x 2 / 4 x 3 / 4 x 1 / 4
In this case, standard normalization is the method proposed by Ko et al. Moreover, we experimented by adding two normalization methods that did not exist before. First, “all reference” is the normalization method in which (3) is applied to the hand. Second, “center reference” uses only each keypoint vector, v, and center vector, c, without a separate reference point. This divides the difference between v and c by distance. Finally, the reference point was compared with the Kim et al. [8] normalization method by fixing it with the right shoulder.
Customized normalization proved to be the most powerful in the experiment. The performance has improved significantly compared to the methods of Kim et al. and Ko et al., which are previous studies. Furthermore, it has proven that it is a new powerful method that can improve performance by performing better in the dataset, KETI, with a relatively large scale frame and in relatively small images, PHOENIX.
All of the methods compared, except our proposed method, have the issue of not properly accounting for the length of human body parts. First, commonly used methods such as Standard, Robust, and Minmax do not account for the fact that people have different body lengths, resulting in poor performance. Secondly, the method used by Kim et al. and the “Center Reference” method, which both use a single reference point, also fail to properly consider the length of each body part. Finally, the “All Reference” method fails to achieve good performance by changing the distribution of hand gestures, which are the most important in sign language. In contrast, customized normalization places a reference point for each body part, allowing the model to understand the length of the human body, and uses min-max normalization for the hand gestures to maintain their distribution, resulting in good performance.

4.2. Effect of SASS

This section shows a difference in performance according to frame selection method experiments. Keypoints were normalized with our experimental normalization method, and everything else was fixed except the frame selection method. We demonstrate that our proposed method is excellent with three comparative experiments. We conduct comparison experiments using only sampling, only augmentation, and both sampling and augmentation methods.
To evaluate the impact of frame sampling, we conduct experiments with different numbers of input frames, N, for each dataset. The KETI dataset is set to N = 50 , which aligns with the results of prior research by Ko et al. For the PHOENIX dataset, N is set to N = 16 , the smallest number of frames among the videos in the training dataset, as no similar methods have been reported in previous studies. In this experiment, our goal is to compare the results of frame sampling only, so we set the number of frames to the minimum among the training videos to eliminate any effects of frame augmentation. The skip sampling method showed the best results for the two datasets in the video sampling comparison experiment. Moreover, the stochastic sampling method differs in performance significantly from random sampling following the uniform distribution. While stochastic sampling is more effective than random sampling because it does not miss key frames, skip sampling was more effective because it extracts key frames by taking into account the gap between the frame and the key frame.
Second, we experimented with the performance change when only frames were augmented. We set the maximum frame value of the training video in the dataset to N and the N of KETI and PHOENIX to 426 and 475, respectively. For the same reason as the experiment that proves the effect of sampling, we set the maximum number of frames of training videos so that frames are not lost (i.e., without the sampling process).
We compared and experimented with a randomly augmented method following an even distribution and a stochastic augmented method following our method. This is shown in Table 3. At KETI, the random augmentation was high performance in every way. However, in PHEONIX, the main evaluation index, BLEU-4, was the best in stochastic augmentation. The overall performance was improved when the frame augmentation was performed compared to when only sampling was performed.
Finally, we experimented with a combination of sampling and augmentation. We experiment with a total of 4 cases. In the previous experiment, since the random sampling method among sampling methods had poor results, we experimented by combining the number of cases except random sampling. This is shown in Table 4. Our method, which combines skip sampling and stochastic augmentation, showed excellent performance in KETI, and BLEU-4 showed the highest performance in PHOENIX.

4.3. Additional Experiments

In this section, we experiment with several factors affecting the model. We compare the performance according to the change in the probability set, P r n , and compare it according to N. Moreover, we experiment with how to invert the order of input values, a method of enhancing the translation performance proposed in the previous study of Sutskever et al. [30]. For the first time, we experiment with the performance change according to l p , the length of probability set P r n . We experiment only with l p [ 3 , 5 , 7 , 9 , 11 , 13 , 15 , 17 ] and evaluate only with BLEU-4. This is shown in Table 5. Our methods outperform other methodologies, including the average BLEU value in the PHOENIX dataset, achieving the best performance on both datasets in the test set with l p of 17.
Second, we experiment with comparison according to N. We set and compare N according to the training video’s mean and median frame length. The median number of frames of training videos in these two datasets is 148 and 112 frames, respectively. The comparison accordingly is shown in Table 6.
Finally, we experiment with the effect of encoding by inverting the order of input X. To this end, we reverse the order of frames and encode each video by changing the order of frames to ( V T * , V T 1 * , , V 1 * ) . This is shown in Table 7. These results show that setting N as the average has the best effect and that entering the frame’s original order has the best effect. It is analyzed that the change in the order of the input sentences is meaningless because our input value is the keypoint value of the frame, not the sentence.

4.4. Comparison of Previous Works

In this section, we compared our keypoint and sign language video processing method with a previous study that used the same deep learning model as ours to demonstrate their superiority. In other words, we compared our approach with a study that used an RNN-based Encoder-Decoder model with Attention [13,49]. The KETI dataset was implemented directly and conducted a comparative experiment because we randomly divided it into the training, validation, and test sets. In Ko et al.’s study, additional data were used, and we reproduced and compared the proposed model without adding data, as shown in Table 8.
Moreover, since we present a robust method of experimenting without glosses dataset, the PHOENIX dataset only compares video-to-text translation without glosses. Most recent studies use glosses, so we compare them with Camgoz et al. [4]. This is shown in Table 9.
Our model also demonstrates high performance compared to previous studies in the dataset. Notably, the KETI dataset exhibits the best performance among keypoint based SLT models. Furthermore, the BLEU score was the highest in the PHOENIX dataset. BLEU-4 increased significantly (about 4%p) compared to Sign2Text (Bahdanau Attention), indicating that the ideal performance was improved through our video-processing method even though the same Bahdanau Attention was used.

5. Ablation Study

We conduct ablation studies to analyze the effectiveness of our proposed method. Change in performance based on the presence or absence of our normalization and frame selection techniques was experimented with. This is shown in Table 10.
We compare the use of our normalization methods. If customized normalization method was not used, the standard normalization method was used. Moreover, the number of all cases was experimented with by applying different frame selection methods. We only use the skip sampling method for sampling and the stochastic augmentation method for augmentation. Finally, the SASS method, which is our method combining the two methods, is compared when applied.
This experiment proves that we achieve greater performance when using customized normalization and the greatest performance when using SASS.

6. Conclusions and Future Work

This paper presents a keypoint-based sign language translation (SLT) model that does not utilize glosses and a novel video processing method that is optimal for this model. We introduce a novel normalization method for keypoint vector processing in sign language videos that are based on human joint positions. Additionally, we propose a method for improving the robustness of our model across multiple datasets by determining the appropriate number of input frames for each dataset rather than using a fixed number of frames for all datasets. Inspired by Camgoz et al. [4], we also demonstrate how the SLT problem can be approached as a machine translation problem.
To demonstrate the effectiveness of our method, we conducted experiments on high-resolution (KETI) and low-resolution (RWTH-PHOENIX-Weather-2014 T) sign language video datasets. The results of these experiments showed that our method is robust across datasets of various sizes. Furthermore, our video-processing method can be applied across various tasks beyond SLT that involve video processing.
In future work, we plan to extend our approach to Sign-to-Gloss-to-Text (S2G2T) translation by leveraging our video-processing method and glosses, as suggested by Camgoz et al. [28] and Yin et al. [25]. We further explore using various neural machine translation models to enhance the performance of our model.

Author Contributions

Conceptualization, H.B.; Methodology, Y.K.; Software, Y.K.; Validation, Y.K.; Investigation, H.B.; Resources, H.B.; Data curation, Y.K.; Writing—original draft, Y.K.; Writing—review & editing, H.B.; Visualization, Y.K.; Supervision, H.B. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Incheon National University Research Grant in 2021.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Santos, A.S.; Portes, A.J.F. Perceptions of deaf subjects about communication in Primary Health Care. Rev. Lat. Am. Enferm. 2019, 27. [Google Scholar] [CrossRef] [PubMed]
  2. Tamura, S.; Kawasaki, S. Recognition of sign language motion images. Pattern Recognit. 1988, 21, 343–353. [Google Scholar] [CrossRef]
  3. Cormier, K.; Fox, N.; Woll, B.; Zisserman, A.; Camgöz, N.C.; Bowden, R. Extol: Automatic recognition of british sign language using the bsl corpus. In Proceedings of the 6th Workshop on Sign Language Translation and Avatar Technology (SLTAT) 2019, Hamburg, Germany, 29 September 2019; Universitat Hamburg: Hamburg, Germany, 2019. [Google Scholar]
  4. Camgoz, N.C.; Hadfield, S.; Koller, O.; Ney, H.; Bowden, R. Neural sign language translation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7784–7793. [Google Scholar]
  5. Sutton-Spence, R.; Woll, B. The Linguistics of British Sign Language: An Introduction; Cambridge University Press: Cambridge, UK, 1999. [Google Scholar]
  6. Braem, P.B.; Sutton-Spence, R. The Hands are the Head of the Mouth: The Mouth as Articulator in Sign Languages; Signum Press: Hamburg, Germany, 2001. [Google Scholar]
  7. Ko, S.K.; Kim, C.J.; Jung, H.; Cho, C. Neural sign language translation based on human keypoint estimation. Appl. Sci. 2019, 9, 2683. [Google Scholar] [CrossRef] [Green Version]
  8. Kim, S.; Kim, C.J.; Park, H.M.; Jeong, Y.; Jang, J.Y.; Jung, H. Robust Keypoint Normalization Method for Korean Sign Language Translation using Transformer. In Proceedings of the 2020 International Conference on Information and Communication Technology Convergence (ICTC), Jeju Island, Republic of Korea, 21–23 October 2020; IEEE: New York, NY, USA, 2020; pp. 1303–1305. [Google Scholar]
  9. Park, C.I.; Sohn, C.B. Data augmentation for human keypoint estimation deep learning based sign language translation. Electronics 2020, 9, 1257. [Google Scholar] [CrossRef]
  10. Gan, S.; Yin, Y.; Jiang, Z.; Xie, L.; Lu, S. Skeleton-aware neural sign language translation. In Proceedings of the 29th ACM International Conference on Multimedia, Online, 20–24 October 2021; pp. 4353–4361. [Google Scholar]
  11. Cho, K.; Van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
  12. Sutskever, I.; Vinyals, O.; Le, Q.V. Sequence to sequence learning with neural networks. Adv. Neural Inf. Process. Syst. 2014, 27. [Google Scholar]
  13. Bahdanau, D.; Cho, K.; Bengio, Y. Neural machine translation by jointly learning to align and translate. arXiv 2014, arXiv:1409.0473. [Google Scholar]
  14. Morrissey, S.; Somers, H.; Smith, R.; Gilchrist, S.; Dandapat, S. Building a Sign Language corpus for use in Machine Translation. In Proceedings of the LREC2010 4th Workshop on the Representation and Processing of Sign Languages: Corpora and Sign Language Technologies, Hamburg, Germany, 17–23 May 2010. [Google Scholar]
  15. Schmidt, C.; Koller, O.; Ney, H.; Hoyoux, T.; Piater, J. Using viseme recognition to improve a sign language translation system. In Proceedings of the 10th International Workshop on Spoken Language Translation: Papers, Heidelberg, Germany, 5–6 December 2013. [Google Scholar]
  16. Stein, D.; Schmidt, C.; Ney, H. Analysis, preparation, and optimization of statistical sign language machine translation. Mach. Transl. 2012, 26, 325–357. [Google Scholar] [CrossRef]
  17. Buehler, P.; Zisserman, A.; Everingham, M. Learning sign language by watching TV (using weakly aligned subtitles). In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; IEEE: New York, NY, USA, 2009; pp. 2961–2968. [Google Scholar]
  18. Cooper, H.; Bowden, R. Learning signs from subtitles: A weakly supervised approach to sign language recognition. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; IEEE: New York, NY, USA, 2009; pp. 2568–2574. [Google Scholar]
  19. Cao, Z.; Simon, T.; Wei, S.E.; Sheikh, Y. Realtime multi-person 2d pose estimation using part affinity fields. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7291–7299. [Google Scholar]
  20. Fang, H.S.; Xie, S.; Tai, Y.W.; Lu, C. Rmpe: Regional multi-person pose estimation. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2334–2343. [Google Scholar]
  21. Forster, J.; Schmidt, C.; Hoyoux, T.; Koller, O.; Zelle, U.; Piater, J.; Ney, H. Rwth-phoenix-weather: A large vocabulary sign language recognition and translation corpus. In Proceedings of the Eighth International Conference on Language Resources and Evaluation (LREC’12), Istanbul, Turkey, 21–27 May 2012; pp. 3785–3789. [Google Scholar]
  22. Forster, J.; Schmidt, C.; Koller, O.; Bellgardt, M.; Ney, H. Extensions of the sign language recognition and translation corpus RWTH-PHOENIX-weather. In Proceedings of the Ninth International Conference on Language Resources and Evaluation (LREC’14), Reykjavik, Iceland, 26–31 May 2014; pp. 1911–1916. [Google Scholar]
  23. Othman, A.; Jemni, M. English-asl gloss parallel corpus 2012: Aslg-pc12. In Proceedings of the 5th Workshop on the Representation and Processing of Sign Languages: Interactions between Corpus and Lexicon LREC, Istanbul, Turkey, 21–27 May 2012. [Google Scholar]
  24. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef] [Green Version]
  25. Yin, K.; Read, J. Better sign language translation with STMC-transformer. In Proceedings of the 28th International Conference on Computational Linguistics, Barcelona, Spain, 8–13 December 2020; pp. 5975–5989. [Google Scholar]
  26. Zhou, H.; Zhou, W.; Zhou, Y.; Li, H. Spatial-temporal multi-cue network for continuous sign language recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 13009–13016. [Google Scholar]
  27. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30. [Google Scholar]
  28. Camgoz, N.C.; Koller, O.; Hadfield, S.; Bowden, R. Sign language transformers: Joint end-to-end sign language recognition and translation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10023–10033. [Google Scholar]
  29. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  30. Simonyan, K.; Zisserman, A. Two-stream convolutional networks for action recognition in videos. Adv. Neural Inf. Process. Syst. 2014, 27. [Google Scholar]
  31. Tran, D.; Bourdev, L.; Fergus, R.; Torresani, L.; Paluri, M. Learning spatiotemporal features with 3d convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 4489–4497. [Google Scholar]
  32. Feichtenhofer, C.; Fan, H.; Malik, J.; He, K. Slowfast networks for video recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6202–6211. [Google Scholar]
  33. Gowda, S.N.; Rohrbach, M.; Sevilla-Lara, L. Smart frame selection for action recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, Online, 2–9 February 2021; Volume 35, pp. 1451–1459. [Google Scholar]
  34. Soomro, K.; Zamir, A.R.; Shah, M. UCF101: A dataset of 101 human actions classes from videos in the wild. arXiv 2012, arXiv:1212.0402. [Google Scholar]
  35. Karpathy, A.; Toderici, G.; Shetty, S.; Leung, T.; Sukthankar, R.; Fei-Fei, L. Large-scale video classification with convolutional neural networks. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1725–1732. [Google Scholar]
  36. Ke, Q.; An, S.; Bennamoun, M.; Sohel, F.; Boussaid, F. Skeletonnet: Mining deep part features for 3-d action recognition. IEEE Signal Process. Lett. 2017, 24, 731–735. [Google Scholar] [CrossRef] [Green Version]
  37. Yue-Hei Ng, J.; Hausknecht, M.; Vijayanarasimhan, S.; Vinyals, O.; Monga, R.; Toderici, G. Beyond short snippets: Deep networks for video classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 4694–4702. [Google Scholar]
  38. Hoai, M.; Zisserman, A. Improving human action recognition using score distribution and ranking. In Proceedings of the Asian Conference on Computer Vision, Singapore, 1–5 November 2015; Springer: Cham, Switzerland, 2015; pp. 3–20. [Google Scholar]
  39. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial networks. Commun. ACM 2020, 63, 139–144. [Google Scholar] [CrossRef]
  40. Vondrick, C.; Pirsiavash, H.; Torralba, A. Generating videos with scene dynamics. Adv. Neural Inf. Process. Syst. 2016, 29. [Google Scholar]
  41. Fernando, B.; Gavves, E.; Oramas, J.M.; Ghodrati, A.; Tuytelaars, T. Modeling video evolution for action recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 5378–5387. [Google Scholar]
  42. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. Pytorch: An imperative style, high-performance deep learning library. Adv. Neural Inf. Process. Syst. 2019, 32. [Google Scholar]
  43. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  44. Park, E.L.; Cho, S. KoNLPy: Korean natural language processing in Python. In Proceedings of the Annual Conference on Human and Language Technology. Human and Language Technology, Kaunas, Lithuania, 26–27 September 2014; pp. 133–136. [Google Scholar]
  45. Loper, E.; Bird, S. Nltk: The natural language toolkit. arXiv 2002, arXiv:cs/0205028. [Google Scholar]
  46. Papineni, K.; Roukos, S.; Ward, T.; Zhu, W.J. Bleu: A method for automatic evaluation of machine translation. In Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics, Philadelphia, PA, USA, 7–12 July 2002; pp. 311–318. [Google Scholar]
  47. Lin, C.Y. Rouge: A package for automatic evaluation of summaries. In Proceedings of the Text Summarization Branches Out, Barcelona, Spain, 25–26 July 2004; pp. 74–81. [Google Scholar]
  48. Banerjee, S.; Lavie, A. METEOR: An automatic metric for MT evaluation with improved correlation with human judgments. In Proceedings of the ACL Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization, Ann Arbor, MI, USA, 29 June 2005; pp. 65–72. [Google Scholar]
  49. Luong, M.T.; Pham, H.; Manning, C.D. Effective approaches to attention-based neural machine translation. arXiv 2015, arXiv:1508.04025. [Google Scholar]
Figure 1. An overview of our approach. S2T means sign-language-to-text model.
Figure 1. An overview of our approach. S2T means sign-language-to-text model.
Sensors 23 03231 g001
Figure 2. The Full Architecture of our SLT approach. We extract the keypoint vector from each frame and then go through the process of normalizing features. Subsequently, after adjusting the number of frames, encoding is performed. After tokenizing the text, it is put into the decoder through the embedding process. In this case, GE is GRU-Encoder, GD is GRU-Decoder and X is the encoding vector that enters the model.
Figure 2. The Full Architecture of our SLT approach. We extract the keypoint vector from each frame and then go through the process of normalizing features. Subsequently, after adjusting the number of frames, encoding is performed. After tokenizing the text, it is put into the decoder through the embedding process. In this case, GE is GRU-Encoder, GD is GRU-Decoder and X is the encoding vector that enters the model.
Sensors 23 03231 g002
Figure 3. Location and number of the keypoints. We used 55 keypoints and excluded the keypoints of the lower body. Hand keypoints included both left and right hands. Then, we divided each normalization part by color.
Figure 3. Location and number of the keypoints. We used 55 keypoints and excluded the keypoints of the lower body. Hand keypoints included both left and right hands. Then, we divided each normalization part by color.
Sensors 23 03231 g003
Figure 4. An overview of SASS. This is our frame selection method. If the length T of the video frame is less than a fixed scalar N, a stochastic augmentation method is used, and if T is greater than N, skip sampling is used to match T with the number of N.
Figure 4. An overview of SASS. This is our frame selection method. If the length T of the video frame is less than a fixed scalar N, a stochastic augmentation method is used, and if T is greater than N, skip sampling is used to match T with the number of N.
Sensors 23 03231 g004
Figure 5. Change of probability distribution and kurtosis following n. (Top) Probability distribution of probability set, F n * , according to the change in l p . (Bottom) Kurtosis of F n * according to the change in l p .
Figure 5. Change of probability distribution and kurtosis following n. (Top) Probability distribution of probability set, F n * , according to the change in l p . (Bottom) Kurtosis of F n * according to the change in l p .
Sensors 23 03231 g005
Table 1. List of keypoints. We divided them into four parts except for the hands. The number is the index number of the corresponding part, and the reference point is the body part that provides the reference for each part.
Table 1. List of keypoints. We divided them into four parts except for the hands. The number is the index number of the corresponding part, and the reference point is the body part that provides the reference for each part.
PartReference PointIncluded Number
FaceNose (0)0, 1, 2, 3, 4, 11
Upper BodyNeck (12)5, 6, 12
Left ArmLeft Elbow (7)7, 9
Right ArmRight Elbow (8)8, 10
Table 2. Comparison Of Normalization Method.
Table 2. Comparison Of Normalization Method.
MethodKETIRWTH-PHOENIX-Weather 2014 T
BLEU-4ROUGE-LMETEORBLEU-4ROUGE-LMETEOR
devtestdevtestdevtestdevtestdevtestdevtest
Standard78.9976.4478.8476.2878.8476.2812.087.3624.7423.5726.9325.87
Robust79.1776.2979.0776.2379.0876.189.198.0424.7323.5726.9225.86
MinMax79.7579.0179.7578.979.7578.99.888.0526.0424.8627.4726.11
Kim et al. [8] method73.9572.9173.872.7773.872.7711.6710.4527.226.2227.9227.2
All reference77.9876.3578.0976.2278.0676.2211.169.327.0825.6828.2227.33
Center Reference75.6575.0275.8274.8175.7974.8112.18.2827.6625.6828.826.79
Ours85.384.3985.4884.8585.5882.3112.8113.3124.6424.7225.8625.85
Table 3. (TOP) Comparison of SLT Performance by Sampling Method. (Bottom) Comparison of SLT Performance by Augmentation Method.
Table 3. (TOP) Comparison of SLT Performance by Sampling Method. (Bottom) Comparison of SLT Performance by Augmentation Method.
MethodKETIRWTH-PHOENIX-Weather 2014 T
BLEU-4ROUGE-LMETEORBLEU-4ROUGE-LMETEOR
devtestdevtestdevtestdevtestdevtestdevtest
Random sampling76.0975.1177.1716.6177.0876.457.775.9421.7421.1223.3622.63
Skip Sampling82.481.983.3582.7483.3282.969.527.5525.0722.3526.3623.77
Stochastic Sampling78.978.279.9279.0779.7879.088.196.4421.4521.2922.8222.86
Random Augmentation83.6983.6884.1984.1584.1984.2513.7111.328.1127.3627.5926.89
Stochastic Augmentation82.2583.3383.1980.0680.7480.7513.7412.6528.0827.5926.3625.73
Table 4. Comparison of combinations of sampling and augmentation.
Table 4. Comparison of combinations of sampling and augmentation.
KETIRWTH-PHOENIX-Weather 2014 T
BLEU-4ROUGE-LMETEORBLEU-4ROUGE-LMETEOR
devtestdevtestdevtestdevtestdevtestdevtest
Skip + Random84.8884.2485.2584.7585.2984.8911.6411.527.6327.6329.6928.68
Stochastic + Random83.8583.9284.4484.6584.484.5710.3610.1627.4826.7729.4728.38
Stochastic + Stochastic84.9483.585.3783.9385.4583.959.7110.8327.8827.0430.0128.7
SASS (ours)85.384.3985.4884.8585.5885.0712.8113.3124.6424.7225.8625.85
Table 5. BLEU-4 score by l p .
Table 5. BLEU-4 score by l p .
l p KETIPHOENIX
devtestdevtest
384.3883.111.2511.17
584.4983.2712.2310.48
784.483.9712.4211.74
983.4183.8514.8412.28
1184.3583.4513.4912.03
1384.7883.3112.3112.01
1584.7681.6812.6411.68
1785.384.3912.8113.31
Mean84.4883.3812.7511.84
Table 6. BLEU-4 score by N.
Table 6. BLEU-4 score by N.
NKETIPHOENIX
devtestdevtest
Mean85.384.3912.8113.31
Median85.2282.7812.659.27
Table 7. BLEU-4 score in order of input values.
Table 7. BLEU-4 score in order of input values.
OrderKETIPHOENIX
devtestdevtest
Reverse85.184.312.389.13
Original85.384.3912.8113.31
Table 8. Comparison with other models in KETI dataset.
Table 8. Comparison with other models in KETI dataset.
BLEU-4ROUGE-LMETEOR
devtestdevtestdevtest
Ko et al. [7]77.9676.4477.9976.5677.9976.49
Ours85.384.3985.4884.8585.5885.07
Table 9. Comparison with other models in PHOENIX dataset.
Table 9. Comparison with other models in PHOENIX dataset.
BLEU-4ROUGE-LMETEOR
devtestdevtestdevtest
Sign2text
(Luong) [4]
10.009.0031.831.8--
Sign2Text
(Bahdanau) [4]
9.949.5832.630.7--
Ours12.8113.3124.6424.7225.8625.85
Table 10. Ablation Study of Video-Processing Methods. “Normalization” means our customized normalization method. “Sampling” is our skip sampling method and “Augmentation” is our stochastic augmentation method.
Table 10. Ablation Study of Video-Processing Methods. “Normalization” means our customized normalization method. “Sampling” is our skip sampling method and “Augmentation” is our stochastic augmentation method.
KETIRWTH-PHOENIX-Weather-2014 T
NormalizationSamplingAugmentationBLEU-4ROUGE-LMETEORBLEU-4ROUGE-LMETEOR
ine 76.4476.5676.497.4821.9123.75
77.0077.6277.8111.424.6224.93
77.7878.4278.617.3624.526.21
81.982.7482.967.5522.3523.77
80.0680.7480.7512.6527.5925.73
84.3984.8585.0713.3124.7225.85
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kim, Y.; Baek, H. Preprocessing for Keypoint-Based Sign Language Translation without Glosses. Sensors 2023, 23, 3231. https://doi.org/10.3390/s23063231

AMA Style

Kim Y, Baek H. Preprocessing for Keypoint-Based Sign Language Translation without Glosses. Sensors. 2023; 23(6):3231. https://doi.org/10.3390/s23063231

Chicago/Turabian Style

Kim, Youngmin, and Hyeongboo Baek. 2023. "Preprocessing for Keypoint-Based Sign Language Translation without Glosses" Sensors 23, no. 6: 3231. https://doi.org/10.3390/s23063231

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop