Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (15)

Search Parameters:
Keywords = probabilistic grammar

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
16 pages, 13345 KB  
Article
Amortized Parameter Inference for the Arbitrary-Order Hidden Markov Model
by Sixiang Zhang and Liming Cai
Axioms 2026, 15(4), 289; https://doi.org/10.3390/axioms15040289 - 14 Apr 2026
Viewed by 280
Abstract
The arbitrary-order hidden Markov model (α-HMM) is a nontrivial generalization of the standard HMM, designed to model stochastic processes with higher-order dependences among arbitrarily distant random events. The α-HMM admits an efficient Viterbi-style optimal decoding algorithm, making it feasible to [...] Read more.
The arbitrary-order hidden Markov model (α-HMM) is a nontrivial generalization of the standard HMM, designed to model stochastic processes with higher-order dependences among arbitrarily distant random events. The α-HMM admits an efficient Viterbi-style optimal decoding algorithm, making it feasible to discover higher-order dependences among data objects in observed sequential data. Because the α-HMM exceeds the expressive power of standard HMMs, fixed kth-order HMMs, and stochastic context-free grammars, effective probabilistic parameter estimation approaches are required to translate this theoretical expressiveness of the α-HMM into practical utility. This paper introduces a principled methodology for effective estimation of probabilistic parameters of the α-HMM from observed data. In large-scale sequential datasets, higher-order dependencies can vary widely across instances, so a single global parameter set may be inadequate. Instead, an amortized parameter inference approach is proposed for the α-HMM, in which an input-conditioned parameter estimator is learned from data and used to infer instance-specific parameters for each input instance to the decoding algorithm. Specifically, the neural parameter estimator is trained using a composite learning objective that is partially enabled by the optimal decoding algorithm. The effectiveness of the proposed parameter estimation method is demonstrated through empirical results of the application of the α-HMM in biomolecular structure modeling and prediction. Full article
(This article belongs to the Special Issue Stochastic Modeling and Optimization Techniques)
Show Figures

Figure 1

62 pages, 7579 KB  
Article
Phonological Choices Drive F0 Range Expansion and Lengthening in Bengali and English Infant-Directed Speech
by Kristine M. Yu, Sameer ud Dowla Khan and Megha Sundara
Languages 2026, 11(4), 68; https://doi.org/10.3390/languages11040068 - 1 Apr 2026
Viewed by 544
Abstract
This study builds on a small body of work, all on Japanese, demonstrating how intonational phonology is critical for understanding prosodic modifications in infant-directed speech (IDS) relative to adult-directed speech. We performed similar analyses on simulated infant-directed speech vs. reading of a story [...] Read more.
This study builds on a small body of work, all on Japanese, demonstrating how intonational phonology is critical for understanding prosodic modifications in infant-directed speech (IDS) relative to adult-directed speech. We performed similar analyses on simulated infant-directed speech vs. reading of a story in English and Bengali: two languages that – unlike Japanese – both have stress and do not use fundamental frequency (F0) to signal changes in word-level meaning, but that have two very different intonational grammars. These differences allowed us to disentangle previous hypotheses about intonational exaggeration in IDS being concentrated in a particular part of the melody. We tested hypotheses that state this locus of exaggeration is either at: the final position in the melody (final in the intonational phrase), the most unpredictable part of the melody, or in pragmatically informative tones. Our results support the first hypothesis. We found that the phonological choices of speakers to chunk the story into shorter, larger prosodic constituents drive intonational exaggeration in IDS. This is because the intonational phrase-final position in both languages is the site of greatest pre-boundary lengthening and F0 range expansion. We also demonstrate: (i) quantification of predictability in intonational melodies using probabilistic finite state automaton representations of intonational grammars and (ii) F0 statistical analyses that are robust and scalable to large, naturalistic IDS corpora. Full article
(This article belongs to the Special Issue Advances in the Acquisition of Prosody)
Show Figures

Figure 1

24 pages, 1372 KB  
Article
From Harmony to Probability: The Problem of Identifiability and a Bayesian Inference Perspective on Greek Nominal Stress
by Kosmas Kosmidis, Giorgos Markopoulos and Anthi Revithiadou
Appl. Sci. 2026, 16(1), 374; https://doi.org/10.3390/app16010374 - 29 Dec 2025
Viewed by 481
Abstract
Maximum Entropy and Gradient Harmonic Grammar are well-established grammatical models for the analysis of linguistic phenomena, but we demonstrate that their probabilistic versions are inherently unidentifiable, employing parameters that cannot be uniquely determined from empirical data. Through examining Greek nominal stress patterns, we [...] Read more.
Maximum Entropy and Gradient Harmonic Grammar are well-established grammatical models for the analysis of linguistic phenomena, but we demonstrate that their probabilistic versions are inherently unidentifiable, employing parameters that cannot be uniquely determined from empirical data. Through examining Greek nominal stress patterns, we propose a reparameterization approach that introduces two identifiable, phonologically interpretable, parameters. We employ both point estimation and Bayesian inference to calculate their values. The latter approach yields reliable parameter estimates with quantified uncertainty that point estimation cannot offer. Our findings shed light on lesser-known aspects of Greek stress and offer a robust methodological framework for probabilistic phonological modeling across languages. Full article
Show Figures

Figure 1

17 pages, 583 KB  
Article
Cross-Domain Feature Enhancement-Based Password Guessing Method for Small Samples
by Cheng Liu, Junrong Li, Xiheng Liu, Bo Li, Mengsu Hou, Wei Yu, Yujun Li and Wenjun Liu
Entropy 2025, 27(7), 752; https://doi.org/10.3390/e27070752 - 15 Jul 2025
Viewed by 1095
Abstract
As a crucial component of account protection system evaluation and intrusion detection, the advancement of password guessing technology encounters challenges due to its reliance on password data. In password guessing research, there is a conflict between the traditional models’ need for large training [...] Read more.
As a crucial component of account protection system evaluation and intrusion detection, the advancement of password guessing technology encounters challenges due to its reliance on password data. In password guessing research, there is a conflict between the traditional models’ need for large training samples and the limitations on accessing password data imposed by privacy protection regulations. Consequently, security researchers often struggle with the issue of having a very limited password set from which to guess. This paper introduces a small-sample password guessing technique that enhances cross-domain features. It analyzes the password set using probabilistic context-free grammar (PCFG) to create a list of password structure probabilities and a dictionary of password fragment probabilities, which are then used to generate a password set structure vector. The method calculates the cosine similarity between the small-sample password set B from the target area and publicly leaked password sets Ai using the structure vector, identifying the set Amax with the highest similarity. This set is then utilized as a training set, where the features of the small-sample password set are enhanced by modifying the structure vectors of the training set. The enhanced training set is subsequently employed for PCFG password generation. The paper uses hit rate as the evaluation metric, and Experiment I reveals that the similarity between B and Ai can be reliably measured when the size of B exceeds 150. Experiment II confirms the hypothesis that a higher similarity between Ai and B leads to a greater hit rate of Ai on the test set of B, with potential improvements of up to 32% compared to training with B alone. Experiment III demonstrates that after enhancing the features of Amax, the hit rate for the small-sample password set can increase by as much as 10.52% compared to previous results. This method offers a viable solution for small-sample password guessing without requiring prior knowledge. Full article
(This article belongs to the Section Information Theory, Probability and Statistics)
Show Figures

Figure 1

7 pages, 771 KB  
Proceeding Paper
Dynamic Oral English Assessment System Based on Large Language Models for Learners
by Jiaqi Yu and Hafriza Binti Burhanudeen
Eng. Proc. 2025, 98(1), 32; https://doi.org/10.3390/engproc2025098032 - 7 Jul 2025
Viewed by 1367
Abstract
The rapid development of science and technology enables technological innovations to change the way of English oral learning. Based on the use of a large language model (LLM), we developed a novel dynamic evaluation system for oral English, LLM-DAELSL, which combines daily oral [...] Read more.
The rapid development of science and technology enables technological innovations to change the way of English oral learning. Based on the use of a large language model (LLM), we developed a novel dynamic evaluation system for oral English, LLM-DAELSL, which combines daily oral habits and a textbook outline. The model integrates commonly used vocabulary from everyday social speech and authoritative prior knowledge, such as oral language textbooks. It also combines traditional large-scale semantic models with probabilistic algorithms to serve as an oral assessment tool for undergraduate students majoring in English-related fields in universities. The model provides corrective feedback to effectively enhance the proficiency of English learners through guided training at any time and place. The technological principle of the model involves inputting prior template knowledge into the language model for reverse guidance and utilizing the textbooks provided by China’s Ministry of Education. The model facilitates the practice and evaluation of pronunciation, grammar, vocabulary, and fluency. The six-month tracking results showed that the oral proficiency of the system learners was significantly improved in the four aspects, which provides a reference for other language learning method developments. Full article
Show Figures

Figure 1

19 pages, 357 KB  
Article
Hidden Abstract Stack Markov Models with Learning Process
by Mete Özbaltan
Mathematics 2024, 12(13), 2144; https://doi.org/10.3390/math12132144 - 8 Jul 2024
Cited by 1 | Viewed by 1799
Abstract
We present hidden abstract stack Markov models (HASMMs) with their learning process. The HASMMs we offer carry the more expressive nature of probabilistic context-free grammars (PCFGs) while allowing faster parameter fitting of hidden Markov models (HMMs). Both HMMs and PCFGs are widely utilized [...] Read more.
We present hidden abstract stack Markov models (HASMMs) with their learning process. The HASMMs we offer carry the more expressive nature of probabilistic context-free grammars (PCFGs) while allowing faster parameter fitting of hidden Markov models (HMMs). Both HMMs and PCFGs are widely utilized structured models, offering an effective formalism capable of describing diverse phenomena. PCFGs are better accommodated than HMMs such as for expressing natural language processing; however, HMMs outperform PCFGs for parameter fitting. We extend HMMs towards PCFGs for such applications, by associating each state of an HMM with an abstract stack, which can be thought of as the single-stack alphabet of pushdown automata (PDA). As a result, we leverage the expressive capabilities of PCFGs for such applications while mitigating the cubic complexity of parameter learning in the observation sequence length of PCFGs by adopting the bilinear complexity of HMMs. Full article
(This article belongs to the Section E1: Mathematics and Computer Science)
Show Figures

Figure 1

21 pages, 571 KB  
Article
LPG–PCFG: An Improved Probabilistic Context- Free Grammar to Hit Low-Probability Passwords
by Xiaozhou Guo, Kaijun Tan, Yi Liu, Min Jin and Huaxiang Lu
Sensors 2022, 22(12), 4604; https://doi.org/10.3390/s22124604 - 18 Jun 2022
Cited by 2 | Viewed by 4027
Abstract
With the development of the Internet, information security has attracted more attention. Identity authentication based on password authentication is the first line of defense; however, the password-generation model is widely used in offline password attacks and password strength evaluation. In real attack scenarios, [...] Read more.
With the development of the Internet, information security has attracted more attention. Identity authentication based on password authentication is the first line of defense; however, the password-generation model is widely used in offline password attacks and password strength evaluation. In real attack scenarios, high-probability passwords are easy to enumerate; extremely low-probability passwords usually lack semantic structure and, so, are tough to crack by applying statistical laws in machine learning models, but these passwords with lower probability have a large search space and certain semantic information. Improving the low-probability password hit rate in this interval is of great significance for improving the efficiency of offline attacks. However, obtaining a low-probability password is difficult under the current password-generation model. To solve this problem, we propose a low-probability generator–probabilistic context-free grammar (LPG–PCFG) based on PCFG. LPG–PCFG directionally increases the probability of low-probability passwords in the models’ distribution, which is designed to obtain a degeneration distribution that is friendly for generating low-probability passwords. By using the control variable method to fine-tune the degeneration of LPG–PCFG, we obtained the optimal combination of degeneration parameters. Compared with the non-degeneration PCFG model, LPG–PCFG generates a larger number of hits. When generating 107 and 108 times, the number of hits to low-probability passwords increases by 50.4% and 42.0%, respectively. Full article
(This article belongs to the Special Issue Advances in Artificial Intelligence for Cyber Security)
Show Figures

Figure 1

19 pages, 2365 KB  
Article
Generating Optimized Guessing Candidates toward Better Password Cracking from Multi-Dictionaries Using Relativistic GAN
by Sungyup Nam, Seungho Jeon and Jongsub Moon
Appl. Sci. 2020, 10(20), 7306; https://doi.org/10.3390/app10207306 - 19 Oct 2020
Cited by 12 | Viewed by 8234
Abstract
Despite their well-known weaknesses, passwords are still the de-facto authentication method for most online systems. Due to its importance, password cracking has been vibrantly researched both for offensive and defensive purposes. Hashcat and John the Ripper are the most popular cracking tools, allowing [...] Read more.
Despite their well-known weaknesses, passwords are still the de-facto authentication method for most online systems. Due to its importance, password cracking has been vibrantly researched both for offensive and defensive purposes. Hashcat and John the Ripper are the most popular cracking tools, allowing users to crack millions of passwords in a short time. However, their rule-based cracking has an explicit limitation of depending on password-cracking experts to come up with creative rules. To overcome this limitation, a recent trend has been to apply machine learning techniques to research on password cracking. For instance, state-of-the-art password guessing studies such as PassGAN and rPassGAN adopted a Generative Adversarial Network (GAN) and used it to generate high-quality password guesses without knowledge of password structures. However, compared with the probabilistic context-free grammar (PCFG), rPassGAN shows inferior password cracking performance in some cases. It was also observed that each password cracker has its own cracking space that does not overlap with other models. This observation led us to realize that an optimized candidate dictionary can be made by combining the password candidates generated by multiple password generation models. In this paper, we suggest a deep learning-based approach called REDPACK that addresses the weakness of the cutting-edge cracking tools based on GAN. To this end, REDPACK combines multiple password candidate generator models in an effective way. Our approach uses the discriminator of rPassGAN as the password selector. Then, by collecting passwords selectively, our model achieves a more realistic password candidate dictionary. Also, REDPACK improves password cracking performance by incorporating both the generator and the discriminator of GAN. We evaluated our system on various datasets with password candidates composed of symbols, digits, upper and lowercase letters. The results clearly show that our approach outperforms all existing approaches, including rule-based Hashcat, GAN-based PassGAN, and probability-based PCFG. The proposed model was also able to reduce the number of password candidates by up to 65%, with only 20% cracking performance loss compared to the union set of passwords cracked by multiple-generation models. Full article
(This article belongs to the Special Issue AI for Cybersecurity)
Show Figures

Figure 1

25 pages, 1083 KB  
Article
Claim Consistency Checking Using Soft Logic
by Nouf Bindris, Nello Cristianini and Jonathan Lawry
Mach. Learn. Knowl. Extr. 2020, 2(3), 147-171; https://doi.org/10.3390/make2030009 - 6 Jul 2020
Viewed by 4785
Abstract
Increasing concerns about the prevalence of false information and fake news has led to calls for automated fact-checking systems that are capable of verifying the truthfulness of statements, especially on the internet. Most previous automated fact-checking systems have focused on the use of [...] Read more.
Increasing concerns about the prevalence of false information and fake news has led to calls for automated fact-checking systems that are capable of verifying the truthfulness of statements, especially on the internet. Most previous automated fact-checking systems have focused on the use of grammar rules only for determining the properties of the language used in statements. Here, we demonstrate a novel approach to the fact-checking of natural language text, which uses a combination of all the following techniques: knowledge extraction to establish a knowledge base, logical inference for fact-checking of claims not explicitly mentioned in the text through the verification of the consistency of a set of beliefs with established trusted knowledge, and a re-querying approach that enables continuous learning. The approach that is presented here addresses the limitations of existing automated fact-checking systems via this novel procedure. This procedure is as follows: the approach investigates the consistency of presented facts or claims while using probabilistic soft logic and a Knowledge Base, which is continuously updated through continuous learning strategies. We demonstrate this approach by focusing on the task of checking facts about family-tree relationships against a corpus of web resources concerned with the UK Royal Family. Full article
(This article belongs to the Section Data)
Show Figures

Figure 1

19 pages, 1978 KB  
Article
Recurrent GANs Password Cracker For IoT Password Security Enhancement
by Sungyup Nam, Seungho Jeon, Hongkyo Kim and Jongsub Moon
Sensors 2020, 20(11), 3106; https://doi.org/10.3390/s20113106 - 31 May 2020
Cited by 35 | Viewed by 10585
Abstract
Text-based passwords are a fundamental and popular means of authentication. Password authentication can be simply implemented because it does not require any equipment, unlike biometric authentication, and it relies only on the users’ memory. This reliance on memory is a weakness of passwords, [...] Read more.
Text-based passwords are a fundamental and popular means of authentication. Password authentication can be simply implemented because it does not require any equipment, unlike biometric authentication, and it relies only on the users’ memory. This reliance on memory is a weakness of passwords, and people therefore usually use easy-to-remember passwords, such as “iloveyou1234”. However, these sample passwords are not difficult to crack. The default passwords of IoT also are text-based passwords and are easy to crack. This weakness enables free password cracking tools such as Hashcat and JtR to execute millions of cracking attempts per second. Finally, this weakness creates a security hole in networks by giving hackers access to an IoT device easily. Research has been conducted to better exploit weak passwords to improve password-cracking performance. The Markov model and probabilistic context-free-grammar (PCFG) are representative research results, and PassGAN, which uses generative adversarial networks (GANs), was recently introduced. These advanced password cracking techniques contribute to the development of better password strength checkers. We studied some methods of improving the performance of PassGAN, and developed two approaches for better password cracking: the first was changing the convolutional neural network (CNN)-based improved Wasserstein GAN (IWGAN) cost function to an RNN-based cost function; the second was employing the dual-discriminator GAN structure. In the password cracking performance experiments, our models showed 10–15% better performance than PassGAN. Through additional performance experiments with PCFG, we identified the cracking performance advantages of PassGAN and our models over PCFG. Finally, we prove that our models enhanced password strength estimation through a comparison with zxcvbn. Full article
(This article belongs to the Special Issue Selected papers from WISA 2019)
Show Figures

Figure 1

20 pages, 3039 KB  
Article
Study on Massive-Scale Slow-Hash Recovery Using Unified Probabilistic Context-Free Grammar and Symmetrical Collaborative Prioritization with Parallel Machines
by Tianjun Wu, Yuexiang Yang, Chi Wang and Rui Wang
Symmetry 2019, 11(4), 450; https://doi.org/10.3390/sym11040450 - 1 Apr 2019
Cited by 3 | Viewed by 4603
Abstract
Slow-hash algorithms are proposed to defend against traditional offline password recovery by making the hash function very slow to compute. In this paper, we study the problem of slow-hash recovery on a large scale. We attack the problem by proposing a novel concurrent [...] Read more.
Slow-hash algorithms are proposed to defend against traditional offline password recovery by making the hash function very slow to compute. In this paper, we study the problem of slow-hash recovery on a large scale. We attack the problem by proposing a novel concurrent model that guesses the target password hash by leveraging known passwords from a largest-ever password corpus. Previously proposed password-reused learning models are specifically designed for targeted online guessing for a single hash and thus cannot be efficiently parallelized for massive-scale offline recovery, which is demanded by modern hash-cracking tasks. In particular, because the size of a probabilistic context-free grammar (PCFG for short) model is non-trivial and keeping track of the next most probable password to guess across all global accounts is difficult, we choose clever data structures and only expand transformations as needed to make the attack computationally tractable. Our adoption of max-min heap, which globally ranks weak accounts for both expanding and guessing according to unified PCFGs and allows for concurrent global ranking, significantly increases the hashes can be recovered within limited time. For example, 59.1% accounts in one of our target password list can be found in our source corpus, allowing our solution to recover 20.1% accounts within one week at an average speed of 7200 non-identical passwords cracked per hour, compared to previous solutions such as oclHashcat (using default configuration), which cracks at an average speed of 28 and needs months to recover the same number of accounts with equal computing resources (thus are infeasible for a real-world attacker who would maximize the gain against the cracking cost). This implies an underestimated threat to slow-hash protected password dumps. Our method provides organizations with a better model of offline attackers and helps them better decide the hashing costs of slow-hash algorithms and detect potential vulnerable credentials before hackers do. Full article
Show Figures

Figure 1

25 pages, 1509 KB  
Article
Critical Behavior in Physics and Probabilistic Formal Languages
by Henry W. Lin and Max Tegmark
Entropy 2017, 19(7), 299; https://doi.org/10.3390/e19070299 - 23 Jun 2017
Cited by 71 | Viewed by 14273
Abstract
We show that the mutual information between two symbols, as a function of the number of symbols between the two, decays exponentially in any probabilistic regular grammar, but can decay like a power law for a context-free grammar. This result about formal languages [...] Read more.
We show that the mutual information between two symbols, as a function of the number of symbols between the two, decays exponentially in any probabilistic regular grammar, but can decay like a power law for a context-free grammar. This result about formal languages is closely related to a well-known result in classical statistical mechanics that there are no phase transitions in dimensions fewer than two. It is also related to the emergence of power law correlations in turbulence and cosmological inflation through recursive generative processes. We elucidate these physics connections and comment on potential applications of our results to machine learning tasks like training artificial recurrent neural networks. Along the way, we introduce a useful quantity, which we dub the rational mutual information, and discuss generalizations of our claims involving more complicated Bayesian networks. Full article
(This article belongs to the Section Information Theory, Probability and Statistics)
Show Figures

Figure 1

15 pages, 765 KB  
Article
CoSpa: A Co-training Approach for Spam Review Identification with Support Vector Machine
by Wen Zhang, Chaoqi Bu, Taketoshi Yoshida and Siguang Zhang
Information 2016, 7(1), 12; https://doi.org/10.3390/info7010012 - 9 Mar 2016
Cited by 24 | Viewed by 5718
Abstract
Spam reviews are increasingly appearing on the Internet to promote sales or defame competitors by misleading consumers with deceptive opinions. This paper proposes a co-training approach called CoSpa (Co-training for Spam review identification) to identify spam reviews by two views: one is the [...] Read more.
Spam reviews are increasingly appearing on the Internet to promote sales or defame competitors by misleading consumers with deceptive opinions. This paper proposes a co-training approach called CoSpa (Co-training for Spam review identification) to identify spam reviews by two views: one is the lexical terms derived from the textual content of the reviews and the other is the PCFG (Probabilistic Context-Free Grammars) rules derived from a deep syntax analysis of the reviews. Using SVM (Support Vector Machine) as the base classifier, we develop two strategies, CoSpa-C and CoSpa-U, embedded within the CoSpa approach. The CoSpa-C strategy selects unlabeled reviews classified with the largest confidence to augment the training dataset to retrain the classifier. The CoSpa-U strategy randomly selects unlabeled reviews with a uniform distribution of confidence. Experiments on the spam dataset and the deception dataset demonstrate that both the proposed CoSpa algorithms outperform the traditional SVM with lexical terms and PCFG rules in spam review identification. Moreover, the CoSpa-U strategy outperforms the CoSpa-C strategy when we use the absolute value of decision function of SVM as the confidence. Full article
Show Figures

Figure 1

5 pages, 68 KB  
Obituary
Ray Solomonoff, Founding Father of Algorithmic Information Theory
by Paul M.B. Vitanyi
Algorithms 2010, 3(3), 260-264; https://doi.org/10.3390/a3030260 - 20 Jul 2010
Viewed by 11785
Abstract
Ray J. Solomonoff died on December 7, 2009, in Cambridge, Massachusetts, of complications of a stroke caused by an aneurism in his head. Ray was the first inventor of Algorithmic Information Theory which deals with the shortest effective description length of objects and [...] Read more.
Ray J. Solomonoff died on December 7, 2009, in Cambridge, Massachusetts, of complications of a stroke caused by an aneurism in his head. Ray was the first inventor of Algorithmic Information Theory which deals with the shortest effective description length of objects and is commonly designated by the term “Kolmogorov complexity.” In the 1950s Solomonoff was one of the first researchers to treat probabilistic grammars and the associated languages. He treated probabilistic Artificial Intelligence (AI) when “probabilistic” was unfashionable, and treated questions of machine learning early on. But his greatest contribution is the creation of Algorithmic Information Theory. [...] Full article
12 pages, 729 KB  
Article
Parsing Costs as Predictors of Reading Difficulty: An Evaluation Using the Potsdam Sentence Corpus
by Marisa Ferrara Boston, John Hale, Reinhold Kliegl, Umesh Patil and Shravan Vasishth
J. Eye Mov. Res. 2008, 2(1), 1-12; https://doi.org/10.16910/jemr.2.1.1 - 8 Sep 2008
Cited by 125 | Viewed by 802
Abstract
The surprisal of a word on a probabilistic grammar constitutes a promising complexity metric for human sentence comprehension difficulty. Using two different grammar types, surprisal is shown to have an effect on fixation durations and regression probabilities in a sample of German readers’ [...] Read more.
The surprisal of a word on a probabilistic grammar constitutes a promising complexity metric for human sentence comprehension difficulty. Using two different grammar types, surprisal is shown to have an effect on fixation durations and regression probabilities in a sample of German readers’ eye movements, the Potsdam Sentence Corpus. A linear mixed-effects model was used to quantify the effect of surprisal while taking into account unigram frequency and bigram frequency (transitional probability), word length, and empirically-derived word predictability; the so-called “early” and “late” measures of processing difficulty both showed an effect of surprisal. Surprisal is also shown to have a small but statistically non-significant effect on empirically-derived predictability itself. This work thus demonstrates the importance of including parsing costs as a predictor of comprehension difficulty in models of reading, and suggests that a simple identification of syntactic parsing costs with early measures and late measures with durations of post-syntactic events may be difficult to uphold. Full article
Show Figures

Figure 1

Back to TopTop