Language Acquisition and Understanding

Special Issue Editors


E-Mail Website
Guest Editor
Faculty of Information Science and Technology, Language Media Laboratory, Hokkaido University, Sapporo 060-0814, Japan
Interests: knowledge acquisition; emotions; common sense; ethics; cognition
Special Issues, Collections and Topics in MDPI journals

E-Mail Website
Guest Editor
Faculty of Information Science and Technology, Knowledge-Base Laboratory, Hokkaido University, Sapporo 060-0814, Japan
Interests: cheminformatics; knowledge engineering; information retrieval; natural language processing; design research; design process modeling

Special Issue Information

Dear Colleagues,

The remarkable capabilities of Large Language Models have pushed the boundaries of artificial intelligence, yet they also highlight a fundamental gap between statistical pattern matching and genuine comprehension. This Special Issue seeks to explore the critical relationship between how an AI system learns language (acquisition) and what it truly understands.

We invite novel research that moves beyond scaling data and parameters to address the core mechanisms of language understanding. We are particularly interested in work that draws inspiration from human cognition, such as data-efficient learning inspired by child development, grounded acquisition that links language to perception and action, and the emergence of compositional reasoning.

Topics of interest include, but are not limited to the following:

  • Low-resource and continual language learning;
  • Grounded language acquisition in embodied agents;
  • Emergent communication and symbolic reasoning;
  • The role of interaction and social learning in AI;
  • New benchmarks for evaluating deep understanding over surface fluency.

We encourage interdisciplinary submissions that bridge machine learning, cognitive science, and linguistics to help shape the next generation of AI systems that not only generate language but genuinely comprehend it.

Dr. Michal Ptaszynski
Dr. Rafal Rzepka
Prof. Dr. Masaharu Yoshioka
Guest Editors

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 250 words) can be sent to the Editorial Office for assessment.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Machine Learning and Knowledge Extraction is an international peer-reviewed open access monthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 1800 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Keywords

  • core concepts:
    • language acquisition
    • language understanding
    • meaning representation
    • comprehension vs. generation
    • surface fluency vs. deep understanding
  • learning paradigms & methods:
    • data-efficient learning
    • low-resource language learning
    • continual learning/lifelong learning
    • interactive learning
    • self-supervised learning
    • curriculum learning
  • cognitive & linguistic principles:
    • compositionality/systematicity
    • generalization
    • child language development
    • cognitive science
    • psycholinguistics
    • symbolic reasoning
    • causality in language
  • embodiment & grounding:
    • grounded language learning/language grounding
    • embodied AI/embodied agents
    • perception and language
    • language and action
    • multimodal learning
  • AI architectures & models:
    • large language models (LLMs)
    • foundation models
    • neuro-symbolic AI
    • agent-based models
    • emergent communication
  • evaluation & analysis:
    • evaluation benchmarks
    • probing
    • interpretability/explainability
    • robustness
    • shortcut learning

Benefits of Publishing in a Special Issue

  • Ease of navigation: Grouping papers by topic helps scholars navigate broad scope journals more efficiently.
  • Greater discoverability: Special Issues support the reach and impact of scientific research. Articles in Special Issues are more discoverable and cited more frequently.
  • Expansion of research network: Special Issues facilitate connections among authors, fostering scientific collaborations.
  • External promotion: Articles in Special Issues are often promoted through the journal's social media, increasing their visibility.
  • Reprint: MDPI Books provides the opportunity to republish successful Special Issues in book format, both online and in print.

Further information on MDPI's Special Issue policies can be found here.

Published Papers (1 paper)

Order results
Result details
Select all
Export citation of selected articles as:

Research

40 pages, 1792 KB  
Article
Why So Meme? A Comparative and Explainable Analysis of Multimodal Hateful Meme Detection
by Nor Saiful Azam Bin Nor Azmi, Michal Ptaszynski, Fumito Masui and Abu Nowhash Chowdhury
Mach. Learn. Knowl. Extr. 2026, 8(2), 50; https://doi.org/10.3390/make8020050 - 21 Feb 2026
Viewed by 200
Abstract
The rise of toxic content, particularly in the form of hateful memes, poses a significant challenge to social media platforms. This paper presents an empirical comparative study of unimodal and multimodal architectures for toxic content detection. Rather than proposing a novel architecture, the [...] Read more.
The rise of toxic content, particularly in the form of hateful memes, poses a significant challenge to social media platforms. This paper presents an empirical comparative study of unimodal and multimodal architectures for toxic content detection. Rather than proposing a novel architecture, the study evaluates the efficacy of a modular Late Fusion framework (RoBERViT) against specialized unimodal baselines (RoBERTa and ViT) and a generalist Large Multimodal (LLaVA). Both unimodal and multimodal configurations across two distinct benchmarks—the imbalanced Innopolis Hateful Memes dataset and the confounder-driven Facebook Hateful Meme dataset—were explored. Beyond quantitative metrics, this study conducts a qualitative analysis using Explainable AI (LIME) and a Large Multimodal Model (LLaVA) to investigate model reasoning. Results demonstrate that the multimodal fusion model consistently outperformed its unimodal counterparts on the Innopolis Hateful Meme dataset, achieving a toxic class F1-score of 0.6439 compared to the text-only score of 0.5794. However, on the Facebook Hateful Meme dataset, text-only models remain competitive, highlighting the “benign confounder” challenge. The qualitative analysis reveals that text remains the dominant modality, with models often relying on surface-level keywords. Notably, the Vision Transformer frequently uses text overlays as a visual proxy for hate, while the LLaVA model struggles with hallucinated toxicity in benign confounder contexts. These findings underscore the persistent challenge of achieving true multimodal understanding in hate speech detection. Full article
(This article belongs to the Special Issue Language Acquisition and Understanding)
Show Figures

Figure 1

Back to TopTop