Deep Learning-Based Automatic Summarization of Chinese Maritime Judgment Documents
Abstract
:1. Introduction
2. Related Work
2.1. Extractive Summarization
2.2. Abstractive Summarization
2.3. Automatic Summarization of Judgment Documents
2.4. Literature Comments
3. Construction of Domain-Specific Lexicon for Maritime Judgment Documents Via Feature Analysis
3.1. Structural and Content Characteristics of Maritime Judgment Documents
3.2. Construction of the Maritime Judgment Lexicon
4. HybridSumm: A Hybrid Summarization Model for Maritime Judgment Documents
4.1. RoBERTa-DCNN-GR: An Extractive Summarization Model
4.1.1. RoBERTa-wwm-ext Pre-Trained Model
4.1.2. Vector Representations of Sentences
- (1)
- Text serialization. Following BERT’s sentence vector generation paradigm, we prepend the [CLS] token and append the [SEP] token to each sentence. The final-layer hidden state of the [CLS] token is extracted as the sentence vector. Subsequently, the text is tokenized into subwords using RoBERTa-wwm-ext’s WordPiece tokenizer, mapped to token ID sequences via its pre-trained vocabulary, and processed with padding/truncation to a fixed length.
- (2)
- Token embeddings. Each subword token is converted into a vector representation through a pre-trained 768-dimensional embedding matrix. These vectors are then ordered according to the original token sequence to form the input representation.
- (3)
- Segment embeddings. Segment embeddings are primarily designed to distinguish between multiple sentences in the input text. However, they are scarcely used in RoBERTa-wwm-ext due to pre-training optimizations like dynamic masking. Following the model’s default configuration, all inputs employ all-zero segment embeddings.
- (4)
- Position embeddings. Position embeddings encode sequential information of tokens. These embeddings share the same dimensionality as token embeddings (768 dimensions). We employ learnable absolute positional embeddings, where each position index is associated with a unique trainable vector. During training, these embeddings are optimized through backpropagation to better capture positional dependencies between tokens, thereby enhancing contextual semantic modeling.
4.1.3. Dilated Convolutional Neural Network (DCNN)
- (1)
- In the CNN architecture, each of the three convolutional layers employs a 3 × 3 kernel, where the output features are computed from a local 3 × 3 region of the previous layer. By stacking these three layers, the final receptive field expands to 7 (calculated as ), enabling the model to capture contextual information from 7 consecutive characters in the input text. However, this limited receptive field restricts its ability to model long-range dependencies in long sequences.
- (2)
- In the DCNN architecture, the first layer uses a 3 × 3 kernel with a dilation rate of 1 (equivalent to standard convolution), yielding a receptive field of 3. The second layer applies a dilation rate of 2, effectively covering a 5 × 5 region (calculated as ). The third layer further scales the dilation rate to 4, expanding coverage to a 9 × 9 region (calculated as ). By stacking these layers, DCNN achieves a final receptive field of 15 (calculated as ), enabling the capture of global dependencies across 15 characters in the input text.
- (3)
- DCNN expands the receptive field by introducing dilated spacing between kernel elements instead of increasing kernel size, and it maintains the same number of parameters as CNN (with nine weights per layer). This design enables the efficient capture of long-range features, allowing DCNN to significantly outperform CNN in processing long texts, such as summarizing maritime judgment documents.
4.1.4. Gated Linear Unit (GLU) and Residual Mechanism
4.2. Nezha-UniLM-CC: An Abstractive Summarization Model
4.2.1. Nezha Pre-Trained Model
- (1)
- Functional relative positional encoding. To address the limitations of BERT’s absolute positional encoding in handling long sequences (e.g., exceeding the 512-token length constraints during pre-training), Nezha introduces a novel sinusoidal-based relative positional encoding mechanism. By mapping relative positional relationships between tokens into a fixed-dimensional vector space using periodic sinusoidal functions—where each dimension corresponds to a specific sinusoidal frequency—the model dynamically generates positional embeddings. This approach not only overcomes sequence length limitations, but also effectively captures long-range dependencies through its periodic characteristics, significantly improving the modeling capability for ultra-long texts.
- (2)
- Mixed-precision training acceleration. To enhance training efficiency, Nezha employs an FP16/FP32 mixed-precision training strategy [68]. This method maintains FP32-format master weights as the parameter baseline while converting them to FP16 for forward and backward computations. Gradients are then updated back to the FP32 master weights. By leveraging the computational advantages of GPU tensor cores for half-precision operations while ensuring numerical stability, this strategy achieves a 1.5–3x training speed improvement.
- (3)
- Large-scale distributed training optimization. To mitigate the performance degradation of traditional optimizers under large-batch training, Nezha pioneers the use of the LAMB (layer-wise adaptive moments optimizer for batch training) optimizer. LAMB enables stable convergence even when scaling batch sizes to tens of thousands through layer-wise adaptive learning rate adjustments, providing critical technical support for distributed training of ultra-large language models.
4.2.2. UniLM Generative Model
4.2.3. Decoding Optimization Strategies
- (1)
- Copy Mechanism
- (2)
- Coverage Mechanism
- (3)
- Beam Search Optimization
4.3. HybridSumm: A Hybrid Summarization Model for Maritime Judgment Documents
5. Experiment and Discussion
5.1. Experimental Dataset
5.2. Measurement for Evaluation
5.2.1. Single-Document Summary Evaluation Metrics
5.2.2. Multi-Document Summary Evaluation Metrics
5.3. Results and Analysis
5.3.1. Validation of Maritime Judgment Lexicon
- (1)
- RoBERTa-DCNN-GR exhibits the lowest dependence on the maritime judgment lexicon, primarily due to its importance-driven extraction strategy. By directly selecting key sentences from the source text, this model prioritizes high-importance content, even if technical terms are incompletely identified. This approach inherently reduces reliance on precise term segmentation, thereby mitigating the impact of segmentation errors. Consequently, the exclusion of the lexicon results in only limited performance degradation.
- (2)
- Nezha-UniLM-CC shows the strongest dependence on the lexicon, directly related to its “pure generative” architecture design. As an abstractive summarization model, it requires explicit access to domain-specific terminology through a lexicon; otherwise, it is prone to producing ambiguous terms or incorrect citations of laws and regulations. The integration of the lexicon enables the model’s pointer–generator network to copy standardized legal terms during decoding, ensuring precise expression of legal concepts. Furthermore, the predefined keywords of laws and regulations in the lexicon standardize the citation logic of legal bases, enabling the model to accurately cite valid legal sources. Consequently, the maritime judgment lexicon has the greatest impact on Nezha-UniLM-CC’s performance.
- (3)
- HybridSumm demonstrates intermediate dependence on the lexicon, primarily due to its “extraction–abstraction” hybrid architecture. The extraction stage preserves key sentences containing domain terms, while the abstraction stage reorganizes content for fluency. Without the lexicon, the abstraction stage still encounters term precision challenges; however, the retained sentences from the extraction stage partially mitigate this issues, resulting in performance declines between purely extractive and purely abstractive models.
- (4)
- The maritime judgment lexicon has enhanced the performance of all three models—RoBERT-DCNN-GR, Nezha-UniLM-CC, and HybridSumm—to varying degrees, confirming the non-negligible role of domain-specific terminology in maritime legal summarization. By explicitly encoding professional terms and their semantic relationships, the lexicon provides interpretable domain-specific prior constraints for automatic summarization models, thereby improving the accuracy, coherence, and judicial logical validity of generated summaries.
5.3.2. Validation of the HybridSumm Model
- (1)
- M1: the LexRank algorithm [14].
- (2)
- M2: the Word2Vec-based LexRank algorithm. In contrast to M1, which employs TF-IDF-based VSM for sentence vectorization, M2 constructs contextualized sentence representations through Word2Vec embeddings. Consider a sentence s with q feature terms. Let (1 ≤ i ≤ q) represent the d-dimensional embedding of the i-th term. The corresponding sentence vector is computed as the arithmetic mean of all q term embeddings, as shown in Equation (18). The embeddings are generated by a 300-dimensional Word2Vec model trained on a 9.86 GB Chinese Wikipedia corpus using the Gensim library.
- (3)
- M3: the MacBERT-attention algorithm. MacBERT [72], developed by HFL in November 2020, is a BERT-based Chinese pre-trained language model. In this paper, the base version of the MacBERT model [73] with a multi-head attention mechanism is employed to summarize maritime judgment documents. Specifically, based on the attention matrices produced by each head in the last layer of MacBERT’s multi-head attention mechanism, the average attention scores of all heads for each token within the input sequence are computed, with the score of the [CLS] token taken as the sentence score. Subsequently, the top-scoring sentences are extracted to form a summary.
- (4)
- M4: the RoBERTa-DCNN-GR algorithm, which is based on the maritime judgment lexicon, as proposed in Section 4.1.
- (5)
- M5: the Chinese BART model, a Chinese-specific linguistically enhanced adaptation of the BART model [74] originally developed by Facebook AI in 2019. By combining BERT’s bidirectional contextual encoding with GPT-style autoregressive decoding, it achieves exceptional performance in both text generation and comprehension tasks. In this paper, we employ the base version of the Chinese BART [75], fine-tuned on a dataset of 1500 documents, to generate summaries of maritime judgment documents.
- (6)
- M6: the Nezha-UniLM-CC algorithm, which is based on the maritime judgment lexicon, as proposed in Section 4.2.
- (7)
- M7: the HybridSumm algorithm, which is based on the maritime judgment lexicon, as proposed in Section 4.3. Unlike M6 that takes raw maritime judgment documents as input, M7 uses the pre-summary generated by M4 as the input for M6, and then utilizes M6 to produce the final summary.
- (1)
- Among the three groups of summarization methods, the hybrid method exhibits the best performance, followed by the abstractive methods, while the extractive methods rank last. This is mainly due to the differences in their capabilities for deep semantic understanding and information restructuring mechanisms. The hybrid summarization method integrates extraction and generation techniques through a two-stage processing framework (first extracting key sentences, then performing semantic reconstruction) and dynamic balancing mechanisms (e.g., attention mechanisms). This ensures the accuracy of facts while improving linguistic fluency, thus producing the highest quality summaries. The abstractive summarization methods, though capable of transcending text structures to generate new sentences through semantic reorganization, are constrained by semantic drift risks (e.g., long-range dependency bias of the Transformer decoders) and high data dependency, resulting in suboptimal performance compared to the hybrid method. The extractive summarization methods, confined to text structures, face information density bottlenecks and logical discontinuity issues, ultimately exhibiting the weakest adaptability and lowest metrics, and thus ranked last among the three groups of methods.
- (2)
- Among the four extractive summarization methods, M4 shows significant superiority over M1, M2, and M3. This result strongly proves the effectiveness of our proposed extractive method M4. Specifically, M4 employs the RoBERTa-www-ext model to obtain sentence embeddings, utilizes stacked DCNNs for sentence classification, and further optimizes the architecture through GLUs and residual mechanisms. This innovative combination enables the model to comprehensively capture global contextual semantics in long-text maritime judgment documents, thus generating high-quality summaries.
- (3)
- Between the two abstractive summarization methods, M6 shows moderately superior performance compared to M5, with AF1-1, AF1-2, and AF1-L metrics improving by 1.77%, 2.59%, and 2.69%, respectively. This result validates the effectiveness of our proposed abstractive method M6 for Chinese maritime judgment document summarization. Specifically, by integrating the Nezha model, the UniLM model, the copy mechanism from a pointer–generator network, the coverage mechanism, and beam search, M6 can effectively capture domain-specific semantic patterns while maintaining structural coherence, thus exhibiting good performance.
- (4)
- The hybrid summarization method M7, which integrates extractive method M4 and abstractive method M6, outperforms both individual methods. Quantitative analysis shows that M7 achieves improvements in AF1-1, AF1-2, and AF1-L by 16.07%, 19.76%, and 16.52%, respectively, compared to M4, and by 10.48%, 15.07%, and 11.65%, respectively, compared to M6. This demonstrates that in the task of maritime judgment document summarization, using the extractive summaries generated by M4 as the input to M6 can effectively overcome the limitations of individual methods, and significantly enhance summary quality.
- (5)
- Among the seven summarization methods (M1–M7), the hybrid method M7 proposed in this paper demonstrates superior performance in generating summaries for Chinese maritime judgment documents. This result strongly validates the effectiveness of M7 and highlights its exceptional capability in the task of maritime judgment document summarization.
- (1)
- Compared to M1, M2 achieves improvements of 4.00%, 4.75%, and 4.78% in AF1-1, AF1-2, and AF1-L metrics, respectively. This demonstrates that the Word2Vec model, which accounts for semantic relationships and syntactic dependencies between terms, provides superior text representation compared to traditional VSM. By more accurately capturing inter-sentence semantic similarities, the enhanced LexRank algorithm empowered by Word2Vec significantly improves summary quality in maritime judgment document summarization tasks.
- (2)
- The summaries generated by M3 achieve improvements of 2.44%, 4.66%, and 4.65% in AF1-1, AF1-2, and AF1-L metrics compared to M1, while exhibiting decreases of 1.56%, 0.09%, and 0.13% relative to M2. The reason for M3’s superior performance over M1 mainly lies in MacBERT’s advanced linguistic comprehension and context-aware representation capabilities. Compared to VSM, MacBERT captures richer semantic patterns—including complex inter-word semantic relationships and contextual dependencies—enabling more precise text interpretation and key information extraction during summarization. As for the reason why M3 underperforms M2, we consider that it mainly results from MacBERT’s input length constraint (maximum 512 tokens). When processing long-text maritime judgment documents, this limitation leads to partial text truncation, hindering the model’s ability to resolve long-range dependencies and consequently impairing its capacity to capture key document information.
6. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Luhn, H.P. The automatic creation of literature abstracts. IBM J. Res. Dev. 1958, 2, 159–165. [Google Scholar] [CrossRef]
- Gambhir, M.; Gupta, V. Recent automatic text summarization techniques: A survey. Artif. Intell. Rev. 2017, 47, 1–66. [Google Scholar] [CrossRef]
- Liu, J.; Zou, Y. A review of automatic text summarization research in recent 70 years. Inf. Sci. 2017, 35, 154–161. (In Chinese) [Google Scholar]
- Ishikawa, K.; Ando, S.; Okumura, A. Hybrid text summarization method based on the TF method and the LEAD method. In Proceedings of the Second NTCIR Workshop Meeting on Evaluation of Chinese & Japanese Text Retrieval and Text Summarization, Tokyo, Japan, 7–9 March 2001; pp. 325–330. [Google Scholar]
- Edmundson, H.P. New methods in automatic extracting. J. ACM 1969, 16, 264–285. [Google Scholar] [CrossRef]
- Lloret, E.; Palomar, M. A gradual combination of features for building automatic summarisation systems. In Proceedings of the 12th International Conference on Text, Speech and Dialogue, Pilsen, Czech Republic, 13–17 September 2009; pp. 16–23. [Google Scholar]
- PadmaLahari, E.; Kumar, D.V.N.S.; Prasad, S. Automatic text summarization with statistical and linguistic features using successive thresholds. In Proceedings of the 2014 IEEE International Conference on Advanced Communications, Control and Computing Technologies, Ramanathapuram, India, 8–10 May 2014; pp. 1519–1524. [Google Scholar]
- Luo, F.; Wang, J.; He, D. Research on automatic text summarization combining topic feature. Appl. Res. Comput. 2021, 38, 129–133. (In Chinese) [Google Scholar]
- Gong, Y.; Liu, X. Generic text summarization using relevance measure and latent semantic analysis. In Proceedings of the 24th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, New Orleans, LA, USA, 9–13 September 2001; pp. 19–25. [Google Scholar]
- Kar, M.; Nunes, S.; Ribeiro, C. Summarization of changes in dynamic text collections using Latent Dirichlet Allocation model. Inf. Process. Manag. 2015, 51, 809–833. [Google Scholar] [CrossRef]
- Page, L.; Brin, S.; Motwani, R.; Winograd, T. The PageRank Citation Ranking: Bringing Order to the Web; Technical Report; Stanford Infolab: Stanford, CA, USA, 1998. [Google Scholar]
- Zong, Q.; Xia, R.; Zhang, J. Text Data Mining; Springer Nature Singapore Pte Ltd.: Singapore, 2021; pp. 287–299. [Google Scholar]
- Mihalcea, R.; Tarau, P. TextRank: Bringing order into texts. In Proceedings of the 2004 Conference on Empirical Methods in Natural Language Processing, Barcelona, Spain, 25–26 July 2004; pp. 404–411. [Google Scholar]
- Erkan, G.; Radev, D.R. Lexrank: Graph-based lexical centrality as salience in text summarization. J. Artif. Intell. Res. 2004, 22, 457–479. [Google Scholar] [CrossRef]
- Huang, B.; Liu, C. Chinese automatic text summarization based on weighted TextRank. Appl. Res. Comput. 2020, 37, 407–410. (In Chinese) [Google Scholar]
- Liu, H.; Zhang, Y. Chinese Single Document Summarization Based on LexRank. J. Ordnance Equip. Eng. 2017, 38, 85–89. (In Chinese) [Google Scholar]
- AL-Khassawneh, Y.A.; Salim, N.; Jarrah, M. Improving triangle-graph based text summarization using hybrid similarity function. Indian J. Sci. Technol. 2017, 10, 1–15. [Google Scholar] [CrossRef]
- Carbonell, J.; Goldstein, J. The use of MMR, diversity-based reranking for reordering documents and producing summaries. In Proceedings of the 21st Annual ACM SIGIR International Conference on Research and Development in Information Retrieval, Melbourne, Australia, 24–28 August 1998; pp. 335–336. [Google Scholar]
- Pozzi, A.; Barbierato, E.; Toti, D. Cryptoblend: An AI-powered tool for aggregation and summarization of cryptocurrency news. Informatics 2023, 10, 5. [Google Scholar] [CrossRef]
- Yin, W.; Pei, Y. Optimizing sentence modeling and selection for document summarization. Proceedings of 24th International Joint Conference on Artificial Intelligence, Buenos Aires, Argentina, 25–31 July 2015; pp. 1383–1389. [Google Scholar]
- Nenkova, A.; McKeown, K. A Survey of Text Summarization Techniques. In Mining Text Data; Aggarwal, C.C., Zhai, C., Eds.; Springer Science+Business Media: New York, NY, USA, 2012; pp. 43–76. [Google Scholar]
- Kupiec, J.; Pedersen, J.; Chen, F. A trainable document summarizer. In Proceedings of the 18th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, Seattle, WA, USA, 9–13 July 1995; pp. 68–73. [Google Scholar]
- Zhang, J.; Chan, H.Y.; Fung, P.; Cao, L. A comparative study on speech summarization of broadcast news and lecture speech. Proceedings of 8th Annual Conference of the International Speech Communication Association, Antwerp, Belgium, 27–31 August 2007; pp. 2781–2784. [Google Scholar]
- Conroy, J.; O’leary, D. Text summarization via hidden Markov models. In Proceedings of the 24th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, New Orleans, LA, USA, 9–13 September 2001; pp. 406–407. [Google Scholar]
- Osborne, M. Using maximum entropy for sentence extraction. In Proceedings of the ACL-02 Workshop on Automatic Summarization, Phildadelphia, PA, USA, 11–12 July 2002; pp. 1–8. [Google Scholar]
- Chali, Y.; Hasan, S.A. Query-focused multi-document summarization: Automatic data annotations and supervised learning approaches. Nat. Lang. Eng. 2012, 18, 109–145. [Google Scholar] [CrossRef]
- Louis, A.; Joshi, A.; Nenkova, A. Discourse indicators for content selection in summarization. In Proceedings of the 11th Annual Meeting of the Special Interest Group on Discourse and Dialogue, Tokyo, Japan, 24–25 September 2010; pp. 147–156. [Google Scholar]
- Nallapati, R.; Zhai, F.; Zhou, B. SummaRuNNer: A recurrent neural network based sequence model for extractive summarization of document. In Proceedings of the 31st AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; pp. 3075–3081. [Google Scholar]
- Rachabathuni, P.K. A survey on abstractive summarization techniques. In Proceedings of the 2017 International Conference on Inventive Computing and Informatics (ICICI), Coimbatore, India, 23–24 November 2017; pp. 762–765. [Google Scholar]
- Belkebir, R.; Guessoum, A. Concept generalization and fusion for abstractive sentence generation. Expert Syst. Appl. 2016, 53, 43–56. [Google Scholar] [CrossRef]
- Nayeem, M.T.; Chali, Y. Paraphrastic fusion for abstractive multi-sentence compression generation. In Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, Singapore, 6–10 November 2017; pp. 2223–2226. [Google Scholar]
- Embar, V.R.; Deshpande, S.R.; Vaishnavi, A.K.; Jain, V.; Kallimani, J.S. sArAmsha—A Kannada abstractive summarizer. In Proceedings of the 2013 International Conference on Advances in Computing, Communications and Informatics (ICACCI), Mysore, India, 22–25 August 2013; pp. 540–544. [Google Scholar]
- Kurisinkel, L.J.; Zhang, Y.; Varma, V. Abstractive multi-document summarization by partial tree extraction, recombination and linearization. In Proceedings of the 8th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), Taipei, Taiwan, 27 November–1 December 2017; pp. 812–821. [Google Scholar]
- Khan, A.; Salim, N.; Farman, H.; Khan, M.; Jan, B.; Ahmad, A.; Paul, A. Abstractive text summarization based on improved semantic graph approach. Int. J. Parallel Prog. 2018, 46, 992–1016. [Google Scholar] [CrossRef]
- Cho, K.; Van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using RNN encoder-decoder for statistical machine translation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; pp. 1724–1734. [Google Scholar]
- Sutskever, I.; Vinyals, O.; Le, Q.V. Sequence to sequence learning with neural networks. In Proceedings of the 27th International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; Volume 2, pp. 3104–3112. [Google Scholar]
- Rush, A.M.; Chopra, S.; Weston, J. A neural attention model for abstractive sentence summarization. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, Lisbon, Portugal, 17–21 September 2015; pp. 379–389. [Google Scholar]
- Chopra, S.; Auli, M.; Rush, A.M. Abstractive sentence summarization with attentive recurrent neural networks. In Proceedings of the 15th Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, San Diego, CA, USA, 12–17 June 2016; pp. 93–98. [Google Scholar]
- Nallapati, R.; Zhou, B.; dos Santos, C.; Gulcehre, C.; Xiang, B. Abstractive text summarization using sequence-to-sequence RNNs and beyond. In Proceedings of the 20th SIGNLL Conference on Computational Natural Language Learning, Berlin, Germany, 11–12 August 2016; pp. 280–290. [Google Scholar]
- Tan, J.; Wan, X.; Xiao, J. Abstractive document summarization with a graph-based attentional neural model. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Vancouver, BC, Canada, 30 July–4 August 2017; pp. 1171–1181. [Google Scholar]
- Chen, Q.; Zhu, X.; Ling, Z.; Wei, S.; Jiang, H. Distraction-based neural networks for document summarization. In Proceedings of the 25th International Joint Conference on Artificial Intelligence (IJCAI-16), New York, NY, USA, 9–15 July 2016; pp. 2754–2760. [Google Scholar]
- Zhang, J.; Zhao, Y.; Saleh, M.; Liu, P. Pegasus: Pre-training with extracted gap-sentences for abstractive summarization. In Proceedings of the 37th International Conference on Machine Learning, Online, 13–18 July 2020; pp. 11328–11339. [Google Scholar]
- Pozzi, A.; Incremona, A.; Tessera, D.; Toti, D. Mitigating exposure bias in large language model distillation: An imitation learning approach. Neural. Comput. Appl. 2025. [Google Scholar] [CrossRef]
- Radford, A.; Narasimhan, K.; Salimans, T.; Sutskever, I. Improving Language Understanding by Generative Pre-Training. 2018. Available online: https://cdn.openai.com/research-covers/language-unsupervised/language_understanding_paper.pdf (accessed on 16 June 2024).
- Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Minneapolis, MN, USA, 2–7 June 2019; pp. 4171–4186. [Google Scholar]
- Liu, Y. Fine-tune BERT for extractive summarization. arXiv 2019, arXiv:1903.10318. [Google Scholar] [CrossRef]
- Liu, Y.; Lapata, M. Text summarization with pretrained encoders. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, 3–7 November 2019; pp. 3730–3740. [Google Scholar]
- Radford, A.; Wu, J.; Child, R.; Luan, D.; Amodei, D.; Sutskever, I. Language Models are Unsupervised Multitask Learners. 2019. Available online: https://cdn.openai.com/better-language-models/language_models_are_unsupervised_multitask_learners.pdf (accessed on 16 June 2024).
- Kieuvongngam, V.; Tan, B.; Niu, Y. Automatic text summarization of COVID-19 medical research articles using bert and gpt-2. arXiv 2020, arXiv:2006.01997. [Google Scholar] [CrossRef]
- An, Z.; Lai, Y.; Feng, Y. Natural language understanding for legal text: A review. J. Chin. Inf. Process. 2022, 36, 1–11. (In Chinese) [Google Scholar]
- Grover, C.; Hachey, B.; Hughson, I. Automatic summarisation of legal documents. In Proceedings of the 9th International Conference on Artificial Intelligence and Law, Scotland, UK, 24–28 June 2003; pp. 243–251. [Google Scholar]
- Farzindar, A.; Lapalme, G. LetSum, an automatic legal text summarizing system. In Proceedings of the 17th Annual Conference on Legal Knowledge and Information Systems, Berlin, Germany, 8–10 December 2004; pp. 11–18. [Google Scholar]
- Galgani, F.; Compton, P.; Hoffmann, A. Combining different summarization techniques for legal text. In Proceedings of the Workshop on Innovative Hybrid Approaches to the Processing of Textual Data, Avignon, France, 23 April 2012; pp. 115–123. [Google Scholar]
- Duan, X.; Zhang, Y.; Yuan, L. Legal summarization for multi-role debate dialogue via controversy focus mining and multi-task learning. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management, Beijing, China, 3–7 November 2019; pp. 1361–1370. [Google Scholar]
- Bhattacharya, P.; Poddar, S.; Rudra, K.; Ghosh, K.; Ghosh, S. Incorporating domain knowledge for extractive summarization of legal case documents. In Proceedings of the 18th International Conference on Artificial Intelligence and Law, São Paulo, Brazil, 21–25 June 2021; pp. 22–31. [Google Scholar]
- Su, J. “SPACES: Extraction-Abstraction”-Based Long-Text Summarization (LAIC Summary). 2021. Available online: https://spaces.ac.cn/archives/8046 (accessed on 27 July 2024). (In Chinese).
- Zhou, W.; Wang, Z.; Wei, B. Abstractive automatic summarizing model for legal judgment documents. Comput. Sci. 2021, 48, 331–336. (In Chinese) [Google Scholar]
- Anand, D.; Wagh, R. Effective deep learning approaches for summarization of legal texts. J. King Saud Univ.-Com. 2022, 34, 2141–2150. [Google Scholar] [CrossRef]
- Wei, X.; Qin, Y.; Tang, X.; Huang, R.; Chen, Y. Method of abstracting judgement document combined with law. Comput. Eng. Des. 2023, 44, 2844–2850. (In Chinese) [Google Scholar]
- Huang, Y.; Sun, L.; Han, C.; Guo, J. A High-Precision Two-Stage Legal Judgment Summarization. Mathematics 2023, 11, 1320. [Google Scholar] [CrossRef]
- Liu, Y.; Ott, M.; Goyal, N.; Du, J.; Joshi, M.; Chen, D.; Levy, O.; Lewis, M.; Zettlemoyer, L.; Stoyanov, V. Roberta: A robustly optimized bert pretraining approach. arXiv 2019, arXiv:1907.11692. [Google Scholar] [CrossRef]
- Cui, Y.; Che, W.; Liu, T.; Qin, B.; Yang, Z. Pre-training with whole word masking for Chinese bert. IEEE/ACM T. Audio Spe. 2021, 29, 3504–3514. [Google Scholar] [CrossRef]
- Yu, F.; Koltun, V. Multi-scale context aggregation by dilated convolutions. In Proceedings of the 4th International Conference on Learning Representations (ICLR), San Juan, Puerto Rico, 2–4 May 2016. [Google Scholar]
- Dauphin, Y.N.; Fan, A.; Auli, M.; Grangier, D. Language modeling with gated convolutional networks. In Proceedings of the 34th International Conference on Machine Learning—Volume 70, Sydney, Australia, 6–11 August 2017; pp. 933–941. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 29th IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Su, J. DGCNN: A Reading Comprehension-Based Question Answering Model Using CNN. 2018. Available online: https://spaces.ac.cn/archives/5409 (accessed on 29 July 2024). (In Chinese).
- Wei, J.; Ren, X.; Li, X.; Huang, W.; Liao, Y.; Wang, Y.; Lin, J.; Jiang, X.; Chen, X.; Liu, Q. NEZHA: Neural contextualized representation for Chinese language understanding. arXiv 2019, arXiv:1909.00204. [Google Scholar] [CrossRef]
- Micikevicius, P.; Narang, S.; Alben, J.; Diamos, G.; Elsen, E.; Garcia, D.; Ginsburg, B.; Houston, M.; Kuchaiev, O.; Venkatesh, G.; et al. Mixed precision training. In Proceedings of the 6th International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
- Li, D.; Yang, N.; Wang, W.; Wei, F.; Liu, X.; Wang, Y.; Gao, J.; Zhou, M.; Hon, H.W. Unified language model pre-training for natural language understanding and generation. In Proceedings of the 33rd International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; pp. 13063–13075. [Google Scholar]
- See, A.; Liu, P.J.; Manning, C.D. Get to the point: Summarization with pointer-generator networks. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Vancouver, BC, Canada, 30 July–4 August 2017; pp. 1073–1083. [Google Scholar]
- Wiseman, S.; Rush, A.M. Sequence-to-sequence learning as beam-search optimization. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, Austin, TX, USA, 1–5 November 2016; pp. 1296–1306. [Google Scholar]
- Cui, Y.; Che, W.; Liu, T.; Qin, B.; Wang, S.; Hu, G. Revisiting pre-trained models for Chinese natural language processing. In Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2020, Online, 16–20 November 2020; pp. 657–668. [Google Scholar]
- Available online: https://github.com/ymcui/MacBERT (accessed on 27 June 2024).
- Lewis, M.; Liu, Y.; Goyal, N.; Ghazvininejad, M.; Mohamed, A.; Levy, O.; Stoyanov, V.; Zettlemoyer, L. Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 7871–7880. [Google Scholar]
- Available online: https://huggingface.co/uer/bart-base-chinese-cluecorpussmall (accessed on 21 April 2024).
Ranking | Cause of Action | Number of Cases |
---|---|---|
1 | Dispute over contract of freight forwarding by sea or by waters leading to the sea | 9002 |
2 | Disputes over contract of seaman service | 6515 |
3 | Disputes over contract of carriage of goods by sea or by waters leading to the sea | 5892 |
4 | Disputes over liability for personal injury at sea or at waters leading to the sea | 2155 |
5 | Admiralty and maritime disputes | 1935 |
6 | Disputes over contract of supply of ship stores and spares | 1617 |
7 | Disputes over contract of sale and purchase of ship | 1118 |
8 | Disputes over contract of insurance by sea or by waters leading to the sea | 1107 |
9 | Charter party disputes | 1073 |
Maximum | Minimum | Average | Standard Deviation | Median | |
---|---|---|---|---|---|
Judgment documents | 50,104 | 252 | 2784 | 2782 | 2605 |
Gold-standard summaries | 988 | 39 | 273 | 111 | 262 |
Model | AF1-1% | AF1-2% | AF1-L% |
---|---|---|---|
RoBERT-DCNN-GR | 53.88 | 39.88 | 52.14 |
—GMJT | 52.08 | 38.38 | 50.53 |
Nezha-UniLM-CC | 59.47 | 44.57 | 57.01 |
—GMJT | 56.40 | 40.32 | 53.79 |
HybridSumm | 69.95 | 59.64 | 68.66 |
—GMJT | 68.21 | 56.52 | 66.59 |
Parameter | M3 | M4 | M5 | M6 |
---|---|---|---|---|
batch_size | 32 | 64 | 3 | 4 |
epochs | 30 | 20 | 40 | 50 |
max_length | 512 | 512 | 512 | 512 |
hidden_size | 768 | 384 | 768 | 768 |
learning_rate | 5 × 10−5 | 0.001 | 5 × 10−5 | 2 × 10−5 |
warmup_steps | 200 | — | 1000 | — |
dropout_rate | 0.1 | 0.1 | 0.1 | 0.1 |
Type | Method | AF1-1% | AF1-2% | AF1-L% |
---|---|---|---|---|
Extractive | M1 | 34.31 | 16.54 | 28.81 |
M2 | 38.31 | 21.29 | 33.59 | |
M3 | 36.75 | 21.20 | 33.46 | |
M4 | 53.88 | 39.88 | 52.14 | |
Abstractive | M5 | 57.70 | 41.98 | 54.32 |
M6 | 59.47 | 44.57 | 57.01 | |
Hybrid | M7 | 69.95 | 59.64 | 68.66 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Zhang, L.; Li, Y.; Zhang, H. Deep Learning-Based Automatic Summarization of Chinese Maritime Judgment Documents. Appl. Sci. 2025, 15, 5434. https://doi.org/10.3390/app15105434
Zhang L, Li Y, Zhang H. Deep Learning-Based Automatic Summarization of Chinese Maritime Judgment Documents. Applied Sciences. 2025; 15(10):5434. https://doi.org/10.3390/app15105434
Chicago/Turabian StyleZhang, Lin, Yanan Li, and Hongyu Zhang. 2025. "Deep Learning-Based Automatic Summarization of Chinese Maritime Judgment Documents" Applied Sciences 15, no. 10: 5434. https://doi.org/10.3390/app15105434
APA StyleZhang, L., Li, Y., & Zhang, H. (2025). Deep Learning-Based Automatic Summarization of Chinese Maritime Judgment Documents. Applied Sciences, 15(10), 5434. https://doi.org/10.3390/app15105434