Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (22)

Search Parameters:
Keywords = probabilistic reward task

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
23 pages, 10908 KB  
Article
MSF: Multi-Level Spatiotemporal Filtering for Event Denoising via Motion Estimation
by Jiuhe Wang, Kun Yu, Xinghua Xu and Nanliang Shan
Sensors 2026, 26(5), 1437; https://doi.org/10.3390/s26051437 - 25 Feb 2026
Viewed by 387
Abstract
Event cameras provide microsecond-level temporal resolution, low latency, and high dynamic range, enabling robust perception under fast motion and challenging lighting conditions. Nevertheless, event streams are susceptible to background activity, thermal noise, and hot pixels. Their sparse and irregular patterns can corrupt event [...] Read more.
Event cameras provide microsecond-level temporal resolution, low latency, and high dynamic range, enabling robust perception under fast motion and challenging lighting conditions. Nevertheless, event streams are susceptible to background activity, thermal noise, and hot pixels. Their sparse and irregular patterns can corrupt event structures and degrade downstream tasks. We propose MSF, a multi-level spatiotemporal filtering framework that couples motion-compensated aggregation with neighborhood-level verification. In each temporal window, MSF estimates a constant 2D optical flow by maximizing a robust, density-normalized contrast objective on the image of warped events (IWE). We further incorporate polarity–gradient decorrelation to suppress mixed-polarity noise and an explicit peak-suppression regularizer to avoid hot-pixel-induced degeneracy. The motion parameters are optimized via coarse grid initialization followed by gradient-ascent refinement. Based on the estimated motion, MSF performs hierarchical event selection: central events are extracted from high-confidence aggregated regions, local events are recovered through joint spatial–temporal–directional–polarity consistency, and weak border events are identified using a density-normalized probabilistic support model that rewards support from reliable structures while penalizing self-clustering. Experiments on four public benchmarks (DVSNOISE20, DVSMOTION20, DVSCLEAN, and E-MLB) show that MSF consistently improves the Event Structural Ratio (ESR) and outperforms representative baselines across diverse motion regimes and severe low-light noise. Full article
(This article belongs to the Special Issue Event-Driven Vision Sensor Architectures and Application Scenarios)
Show Figures

Figure 1

15 pages, 720 KB  
Article
Sex and Age Differences in Decision-Making Under Risk by Wild Balinese Long-Tailed Macaques (Macaca fascicularis fascicularis): A Field Experimental Study
by Caleb Bunselmeyer, Noëlle Gunst, I Nengah Wandia, Robert J. Williams, Elsa Addessi and Jean-Baptiste Leca
Animals 2026, 16(4), 617; https://doi.org/10.3390/ani16040617 - 15 Feb 2026
Viewed by 710
Abstract
This study examines risky decision-making in a free-ranging population of Balinese long-tailed macaques (Macaca fascicularis fascicularis), addressing gaps in research that have largely focused on captive primates and have rarely considered individual differences by age and sex. Thirty-three macaques of different [...] Read more.
This study examines risky decision-making in a free-ranging population of Balinese long-tailed macaques (Macaca fascicularis fascicularis), addressing gaps in research that have largely focused on captive primates and have rarely considered individual differences by age and sex. Thirty-three macaques of different age–sex classes were tested using a choice task contrasting a guaranteed small reward with a probabilistic larger reward. At the group level, macaques showed no preference for safe or risky options. However, substantial individual variation emerged: some individuals were risk-prone, others risk-averse, and many indifferent. Notably, age and sex interacted in shaping risk preferences. Among males, adults and juveniles were more risk-prone than younger adults, whereas among females, adults were more risk-prone than juveniles. Juveniles also displayed outcome-dependent flexibility, choosing the risky option more often after a previous successful risky choice, consistent with a win–stay strategy. Like in rodents, this pattern may reflect adaptive learning during developmental transitions. Importantly, the observed behavioral differences were not due to misunderstanding of the task, as macaques reliably chose the larger option when outcomes were visible. This pronounced individual variability in primate risk preferences underscore the importance of considering demographic factors when characterizing species-typical risk preferences. Full article
(This article belongs to the Section Human-Animal Interactions, Animal Behaviour and Emotion)
Show Figures

Graphical abstract

25 pages, 2211 KB  
Article
When Demand Uncertainty Occurs in Emergency Supplies Allocation: A Robust DRL Approach
by Weimeng Wang, Junchao Fan, Weiqiao Zhu, Yujing Cai, Yang Yang, Xuanming Zhang, Yingying Yao and Xiaolin Chang
Appl. Sci. 2026, 16(2), 581; https://doi.org/10.3390/app16020581 - 6 Jan 2026
Viewed by 455
Abstract
Emergency supplies allocation is a critical task in post-disaster response, as ineffective or delayed decisions can directly lead to increased human suffering and loss of life. In practice, emergency managers must make rapid allocation decisions over multiple periods under incomplete information and highly [...] Read more.
Emergency supplies allocation is a critical task in post-disaster response, as ineffective or delayed decisions can directly lead to increased human suffering and loss of life. In practice, emergency managers must make rapid allocation decisions over multiple periods under incomplete information and highly unpredictable demand, making robust and adaptive decision support essential. However, existing allocation approaches face several challenges: (1) Those traditional approaches rely heavily on predefined uncertainty sets or probabilistic models, and are inherently static, making them unsuitable for multi-period, dynamically allocation problems; and (2) while reinforcement learning (RL) technique is inherently suitable for dynamic decision-making, most existing RL-base approaches assume fixed demand, making them unable to cope with the non-stationary demand patterns seen in real disasters. To address these challenges, we first establish a multi-period and multi-objective emergency supplies allocation problem with demand uncertainty and then formulate it as a two-player zero-sum Markov game (TZMG). Demand uncertainty is modeled through an adversary rather than predefined uncertainty sets. We then propose RESA, a novel RL framework that uses adversarial training to learn robust allocation policies. In addition, RESA introduces a combinatorial action representation and reward clipping methods to handle high-dimensional allocations and nonlinear objectives. Building on RESA, we develop RESA_PPO by employing proximal policy optimization as its policy optimizer. Experiment results with realistic post-disaster data show that RESA_PPO achieves near-optimal performance, with an average gap of only 3.7% in terms of the objective value of the formulated problem, from the theoretical optimum derived by exact solvers. Moreover, RESA_PPO outperforms all baseline methods, including heuristic and standard RL methods, by at least 5.25% on average. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

18 pages, 1291 KB  
Article
Exploration of Psychosocial Factors in Peruvian Workers: A Quantitative Analysis of Qualitative Categorizations
by Arturo Juárez-García, César Merino-Soto and Javier García-Rivas
Hygiene 2025, 5(4), 43; https://doi.org/10.3390/hygiene5040043 - 30 Sep 2025
Viewed by 1278
Abstract
This study aimed to explore psychosocial factors in a sample of Peruvian workers, examine their convergence with the PROPSIT model, and identify the emergence of new or idiosyncratic psychosocial dimensions. At the same time, the quality and efficiency of the categorization process were [...] Read more.
This study aimed to explore psychosocial factors in a sample of Peruvian workers, examine their convergence with the PROPSIT model, and identify the emergence of new or idiosyncratic psychosocial dimensions. At the same time, the quality and efficiency of the categorization process were evaluated. n = 48 workers were contacted by a non-probabilistic sampling method and asked to fill out a form with open-ended questions that explored negative stressors and positive engaging factors. Some strategies were used to assess the quality and efficiency of the categorization process. The results showed that the quality, speed, and reliability of the categorization procedure were satisfactory, and several categories were aligned with the PROPSIT model and other literature, both in their negative aspects (workload and rhythm, working hours, shifts, etc.) and positive aspects (rewarding tasks, atmosphere of unity, etc.). The emerging new categories were confined to aspects of teamwork and conflict climate, as well as topics such as order, cleanliness, and recreation. These findings underline the need to adapt existing models and instruments to capture idiosyncratic aspects of the Peruvian work environment. In conclusion, this study validated an efficient mixed approach for categorizing psychosocial work factors in Peru, revealing both PROPSIT-aligned and novel context-specific categories, and highlighting the need for culturally adapted tools and broader validation. Full article
Show Figures

Figure 1

21 pages, 5337 KB  
Article
SC-NBTI: A Smart Contract-Based Incentive Mechanism for Federated Knowledge Sharing
by Yuanyuan Zhang, Jingwen Liu, Jingpeng Li, Yuchen Huang, Wang Zhong, Yanru Chen and Liangyin Chen
Sensors 2025, 25(18), 5802; https://doi.org/10.3390/s25185802 - 17 Sep 2025
Cited by 2 | Viewed by 1060
Abstract
With the rapid expansion of digital knowledge platforms and intelligent information systems, organizations and communities are producing a vast number of unstructured knowledge data, including annotated corpora, technical diagrams, collaborative whiteboard content, and domain-specific multimedia archives. However, knowledge sharing across institutions is hindered [...] Read more.
With the rapid expansion of digital knowledge platforms and intelligent information systems, organizations and communities are producing a vast number of unstructured knowledge data, including annotated corpora, technical diagrams, collaborative whiteboard content, and domain-specific multimedia archives. However, knowledge sharing across institutions is hindered by privacy risks, high communication overhead, and fragmented ownership of data. Federated learning promises to overcome these barriers by enabling collaborative model training without exchanging raw knowledge artifacts, but its success depends on motivating data holders to undertake the additional computational and communication costs. Most existing incentive schemes, which are based on non-cooperative game formulations, neglect unstructured interactions and communication efficiency, thereby limiting their applicability in knowledge-driven scenarios. To address these challenges, we introduce SC-NBTI, a smart contract and Nash bargaining-based incentive framework for federated learning in knowledge collaboration environments. We cast the reward allocation problem as a cooperative game, devise a heuristic algorithm to approximate the NP-hard Nash bargaining solution, and integrate a probabilistic gradient sparsification method to trim communication costs while safeguarding privacy. Experiments on the FMNIST image classification task show that SC-NBTI requires fewer training rounds while achieving 5.89% higher accuracy than the DRL-Incentive baseline. Full article
(This article belongs to the Section Internet of Things)
Show Figures

Figure 1

27 pages, 5560 KB  
Article
A Stackelberg Trust-Based Human–Robot Collaboration Framework for Warehouse Picking
by Yang Liu, Fuqiang Guo and Yan Ma
Systems 2025, 13(5), 348; https://doi.org/10.3390/systems13050348 - 3 May 2025
Cited by 5 | Viewed by 2238
Abstract
The warehouse picking process is one of the most critical components of logistics operations. Human–robot collaboration (HRC) is seen as an important trend in warehouse picking, as it combines the strengths of both humans and robots in the picking process. However, in current [...] Read more.
The warehouse picking process is one of the most critical components of logistics operations. Human–robot collaboration (HRC) is seen as an important trend in warehouse picking, as it combines the strengths of both humans and robots in the picking process. However, in current human–robot collaboration frameworks, there is a lack of effective communication between humans and robots, which results in inefficient task execution during the picking process. To address this, this paper considers trust as a communication bridge between humans and robots and proposes the Stackelberg trust-based human–robot collaboration framework for warehouse picking, aiming to achieve efficient and effective human–robot collaborative picking. In this framework, HRC with trust for warehouse picking is defined as the Partially Observable Stochastic Game (POSG) model. We model human fatigue with the logistic function and incorporate its impact on the efficiency reward function of the POSG. Based on the POSG model, belief space is used to assess human trust, and human strategies are formed. An iterative Stackelberg trust strategy generation (ISTSG) algorithm is designed to achieve the optimal long-term collaboration benefits between humans and robots, which is solved by the Bellman equation. The generated human–robot decision profile is formalized as a Partially Observable Markov Decision Process (POMDP), and the properties of human–robot collaboration are specified as PCTL (probabilistic computation tree logic) with rewards, such as efficiency, accuracy, trust, and human fatigue. The probabilistic model checker PRISM is exploited to verify and analyze the corresponding properties of the POMDP. We take the popular human–robot collaboration robot TORU as a case study. The experimental results show that our framework improves the efficiency of human–robot collaboration for warehouse picking and reduces worker fatigue while ensuring the required accuracy of human–robot collaboration. Full article
Show Figures

Figure 1

23 pages, 8280 KB  
Article
From Task Distributions to Expected Paths Lengths Distributions: Value Function Initialization in Sparse Reward Environments for Lifelong Reinforcement Learning
by Soumia Mehimeh and Xianglong Tang
Entropy 2025, 27(4), 367; https://doi.org/10.3390/e27040367 - 30 Mar 2025
Viewed by 779
Abstract
This paper studies value function transfer within reinforcement learning frameworks, focusing on tasks continuously assigned to an agent through a probabilistic distribution. Specifically, we focus on environments characterized by sparse rewards with a terminal goal. Initially, we propose and theoretically demonstrate that the [...] Read more.
This paper studies value function transfer within reinforcement learning frameworks, focusing on tasks continuously assigned to an agent through a probabilistic distribution. Specifically, we focus on environments characterized by sparse rewards with a terminal goal. Initially, we propose and theoretically demonstrate that the distribution of the computed value function from such environments, whether in cases where the goals or the dynamics are changing across tasks, can be reformulated as the distribution of the number of steps to the goal generated by their optimal policies, which we name the expected optimal path length. To test our propositions, we hypothesize that the distribution of the expected optimal path lengths resulting from the task distribution is normal. This claim leads us to propose that if the distribution is normal, then the distribution of the value function follows a log-normal pattern. Leveraging this insight, we introduce “LogQInit” as a novel value function transfer method, based on the properties of log-normality. Finally, we run experiments on a scenario of goals and dynamics distributions, validate our proposition by providing an a dequate analysis of the results, and demonstrate that LogQInit outperforms existing methods of value function initialization, policy transfer, and reward shaping. Full article
(This article belongs to the Section Information Theory, Probability and Statistics)
Show Figures

Figure 1

18 pages, 2027 KB  
Article
Runtime Verification-Based Safe MARL for Optimized Safety Policy Generation for Multi-Robot Systems
by Yang Liu and Jiankun Li
Big Data Cogn. Comput. 2024, 8(5), 49; https://doi.org/10.3390/bdcc8050049 - 16 May 2024
Cited by 5 | Viewed by 3302
Abstract
The intelligent warehouse is a modern logistics management system that uses technologies like the Internet of Things, robots, and artificial intelligence to realize automated management and optimize warehousing operations. The multi-robot system (MRS) is an important carrier for implementing an intelligent warehouse, which [...] Read more.
The intelligent warehouse is a modern logistics management system that uses technologies like the Internet of Things, robots, and artificial intelligence to realize automated management and optimize warehousing operations. The multi-robot system (MRS) is an important carrier for implementing an intelligent warehouse, which completes various tasks in the warehouse through cooperation and coordination between robots. As an extension of reinforcement learning and a kind of swarm intelligence, MARL (multi-agent reinforcement learning) can effectively create the multi-robot systems in intelligent warehouses. However, MARL-based multi-robot systems in intelligent warehouses face serious safety issues, such as collisions, conflicts, and congestion. To deal with these issues, this paper proposes a safe MARL method based on runtime verification, i.e., an optimized safety policy-generation framework, for multi-robot systems in intelligent warehouses. The framework consists of three stages. In the first stage, a runtime model SCMG (safety-constrained Markov Game) is defined for the multi-robot system at runtime in the intelligent warehouse. In the second stage, rPATL (probabilistic alternating-time temporal logic with rewards) is used to express safety properties, and SCMG is cyclically verified and refined through runtime verification (RV) to ensure safety. This stage guarantees the safety of robots’ behaviors before training. In the third stage, the verified SCMG guides SCPO (safety-constrained policy optimization) to obtain an optimized safety policy for robots. Finally, a multi-robot warehouse (RWARE) scenario is used for experimental evaluation. The results show that the policy obtained by our framework is safer than existing frameworks and includes a certain degree of optimization. Full article
(This article belongs to the Special Issue Field Robotics and Artificial Intelligence (AI))
Show Figures

Figure 1

20 pages, 6707 KB  
Article
Towards Multi-Objective Object Push-Grasp Policy Based on Maximum Entropy Deep Reinforcement Learning under Sparse Rewards
by Tengteng Zhang and Hongwei Mo
Entropy 2024, 26(5), 416; https://doi.org/10.3390/e26050416 - 12 May 2024
Viewed by 2781
Abstract
In unstructured environments, robots need to deal with a wide variety of objects with diverse shapes, and often, the instances of these objects are unknown. Traditional methods rely on training with large-scale labeled data, but in environments with continuous and high-dimensional state spaces, [...] Read more.
In unstructured environments, robots need to deal with a wide variety of objects with diverse shapes, and often, the instances of these objects are unknown. Traditional methods rely on training with large-scale labeled data, but in environments with continuous and high-dimensional state spaces, the data become sparse, leading to weak generalization ability of the trained models when transferred to real-world applications. To address this challenge, we present an innovative maximum entropy Deep Q-Network (ME-DQN), which leverages an attention mechanism. The framework solves complex and sparse reward tasks through probabilistic reasoning while eliminating the trouble of adjusting hyper-parameters. This approach aims to merge the robust feature extraction capabilities of Fully Convolutional Networks (FCNs) with the efficient feature selection of the attention mechanism across diverse task scenarios. By integrating an advantage function with the reasoning and decision-making of deep reinforcement learning, ME-DQN propels the frontier of robotic grasping and expands the boundaries of intelligent perception and grasping decision-making in unstructured environments. Our simulations demonstrate a remarkable grasping success rate of 91.6%, while maintaining excellent generalization performance in the real world. Full article
Show Figures

Figure 1

23 pages, 2796 KB  
Article
Evaluating the Influence of Musical and Monetary Rewards on Decision Making through Computational Modelling
by Grigory Kopytin, Marina Ivanova, Maria Herrojo Ruiz and Anna Shestakova
Behav. Sci. 2024, 14(2), 124; https://doi.org/10.3390/bs14020124 - 8 Feb 2024
Viewed by 2318
Abstract
A central question in behavioural neuroscience is how different rewards modulate learning. While the role of monetary rewards is well-studied in decision-making research, the influence of abstract rewards like music remains poorly understood. This study investigated the dissociable effects of these two reward [...] Read more.
A central question in behavioural neuroscience is how different rewards modulate learning. While the role of monetary rewards is well-studied in decision-making research, the influence of abstract rewards like music remains poorly understood. This study investigated the dissociable effects of these two reward types on decision making. Forty participants completed two decision-making tasks, each characterised by probabilistic associations between stimuli and rewards, with probabilities changing over time to reflect environmental volatility. In each task, choices were reinforced either by monetary outcomes (win/lose) or by the endings of musical melodies (consonant/dissonant). We applied the Hierarchical Gaussian Filter, a validated hierarchical Bayesian framework, to model learning under these two conditions. Bayesian statistics provided evidence for similar learning patterns across both reward types, suggesting individuals’ similar adaptability. However, within the musical task, individual preferences for consonance over dissonance explained some aspects of learning. Specifically, correlation analyses indicated that participants more tolerant of dissonance behaved more stochastically in their belief-to-response mappings and were less likely to choose the response associated with the current prediction for a consonant ending, driven by higher volatility estimates. By contrast, participants averse to dissonance showed increased tonic volatility, leading to larger updates in reward tendency beliefs. Full article
Show Figures

Figure 1

12 pages, 1729 KB  
Article
Dynamics Learning Rate Bias in Pigeons: Insights from Reinforcement Learning and Neural Correlates
by Fuli Jin, Lifang Yang, Long Yang, Jiajia Li, Mengmeng Li and Zhigang Shang
Animals 2024, 14(3), 489; https://doi.org/10.3390/ani14030489 - 1 Feb 2024
Cited by 3 | Viewed by 2315
Abstract
Research in reinforcement learning indicates that animals respond differently to positive and negative reward prediction errors, which can be calculated by assuming learning rate bias. Many studies have shown that humans and other animals have learning rate bias during learning, but it is [...] Read more.
Research in reinforcement learning indicates that animals respond differently to positive and negative reward prediction errors, which can be calculated by assuming learning rate bias. Many studies have shown that humans and other animals have learning rate bias during learning, but it is unclear whether and how the bias changes throughout the entire learning process. Here, we recorded the behavior data and the local field potentials (LFPs) in the striatum of five pigeons performing a probabilistic learning task. Reinforcement learning models with and without learning rate biases were used to dynamically fit the pigeons’ choice behavior and estimate the option values. Furthemore, the correlation between the striatal LFPs power and the model-estimated option values was explored. We found that the pigeons’ learning rate bias shifted from negative to positive during the learning process, and the striatal Gamma (31 to 80 Hz) power correlated with the option values modulated by dynamic learning rate bias. In conclusion, our results support the hypothesis that pigeons employ a dynamic learning strategy in the learning process from both behavioral and neural aspects, providing valuable insights into reinforcement learning mechanisms of non-human animals. Full article
(This article belongs to the Section Birds)
Show Figures

Figure 1

13 pages, 983 KB  
Article
Comparison between the Effects of Acute Physical and Psychosocial Stress on Feedback-Based Learning
by Xiao Yang, Brittany Nackley and Bruce H. Friedman
Brain Sci. 2023, 13(8), 1127; https://doi.org/10.3390/brainsci13081127 - 26 Jul 2023
Cited by 3 | Viewed by 2375
Abstract
Stress modulates feedback-based learning, a process that has been implicated in declining mental function in aging and mental disorders. While acute physical and psychosocial stressors have been used interchangeably in studies on feedback-based learning, the two types of stressors involve distinct physiological and [...] Read more.
Stress modulates feedback-based learning, a process that has been implicated in declining mental function in aging and mental disorders. While acute physical and psychosocial stressors have been used interchangeably in studies on feedback-based learning, the two types of stressors involve distinct physiological and psychological processes. Whether the two types of stressors differentially influence feedback processing remains unclear. The present study compared the effects of physical and psychosocial stressors on feedback-based learning. Ninety-six subjects (Mage = 19.11 years; 50 female) completed either a cold pressor task (CPT) or mental arithmetic task (MAT), as the physical or psychosocial stressor, while electrocardiography and blood pressure were measured to assess cardiovascular stress reactivity (CVR). Self-ratings on the emotional valence of the stressors were also obtained. A probabilistic learning task was given prior to and after the stressors. Accuracy in selecting positive (Go accuracy) and avoiding negative stimuli (No-go accuracy) were recorded as learning outcomes. Repeated measures ANOVA and multiple regressions were used to compare the effects of two stressors and examine the effects of CVR and valence on the learning outcomes. The results showed that although the effects of CPT and MAT on feedback processing were not different, CVR and valence influenced Go and No-go accuracy, respectively. The results suggest that stress-modulated feedback-based learning involves multiple pathways and underscore the link between CVR and reward sensitivity. The findings have clinical implications and may contribute to a better understanding of human behavioral systems. Full article
(This article belongs to the Section Behavioral Neuroscience)
Show Figures

Figure 1

11 pages, 573 KB  
Brief Report
Reward Behavior Disengagement, a Neuroeconomic Model-Based Objective Measure of Reward Pathology in Depression: Findings from the EMBARC Trial
by Michael A. Giles, Crystal M. Cooper, Manish K. Jha, Cherise R. Chin Fatt, Diego A. Pizzagalli, Taryn L. Mayes, Christian A. Webb, Tracy L. Greer, Amit Etkin, Joseph M. Trombello, Henry W. Chase, Mary L. Phillips, Melvin G. McInnis, Thomas Carmody, Phillip Adams, Ramin V. Parsey, Patrick J. McGrath, Myrna Weissman, Benji T. Kurian, Maurizio Fava and Madhukar H. Trivediadd Show full author list remove Hide full author list
Behav. Sci. 2023, 13(8), 619; https://doi.org/10.3390/bs13080619 - 25 Jul 2023
Cited by 1 | Viewed by 4350
Abstract
The probabilistic reward task (PRT) has identified reward learning impairments in those with major depressive disorder (MDD), as well as anhedonia-specific reward learning impairments. However, attempts to validate the anhedonia-specific impairments have produced inconsistent findings. Thus, we seek to determine whether the Reward [...] Read more.
The probabilistic reward task (PRT) has identified reward learning impairments in those with major depressive disorder (MDD), as well as anhedonia-specific reward learning impairments. However, attempts to validate the anhedonia-specific impairments have produced inconsistent findings. Thus, we seek to determine whether the Reward Behavior Disengagement (RBD), our proposed economic augmentation of PRT, differs between MDD participants and controls, and whether there is a level at which RBD is high enough for depressed participants to be considered objectively disengaged. Data were gathered as part of the Establishing Moderators and Biosignatures of Antidepressant Response in Clinical Care (EMBARC) study, a double-blind, placebo-controlled clinical trial of antidepressant response. Participants included 195 individuals with moderate to severe MDD (Quick Inventory of Depressive Symptomatology (QIDS–SR) score ≥ 15), not in treatment for depression, and with complete PRT data. Healthy controls (n = 40) had no history of psychiatric illness, a QIDS–SR score < 8, and complete PRT data. Participants with MDD were treated with sertraline or placebo for 8 weeks (stage I of the EMBARC trial). RBD was applied to PRT data using discriminant analysis, and classified MDD participants as reward task engaged (n = 137) or reward task disengaged (n = 58), relative to controls. Reward task engaged/disengaged groups were compared on sociodemographic features, reward–behavior, and sertraline/placebo response (Hamilton Depression Rating Scale scores). Reward task disengaged MDD participants responded only to sertraline, whereas those who were reward task engaged responded to sertraline and placebo (F(1293) = 4.33, p = 0.038). Reward task engaged/disengaged groups did not differ otherwise. RBD was predictive of reward impairment in depressed patients and may have clinical utility in identifying patients who will benefit from antidepressants. Full article
Show Figures

Figure 1

25 pages, 32503 KB  
Article
Real-Time 6-DOF Pose Estimation of Known Geometries in Point Cloud Data
by Vedant Bhandari, Tyson Govan Phillips and Peter Ross McAree
Sensors 2023, 23(6), 3085; https://doi.org/10.3390/s23063085 - 13 Mar 2023
Cited by 6 | Viewed by 6050
Abstract
The task of tracking the pose of an object with a known geometry from point cloud measurements arises in robot perception. It calls for a solution that is both accurate and robust, and can be computed at a rate that aligns with the [...] Read more.
The task of tracking the pose of an object with a known geometry from point cloud measurements arises in robot perception. It calls for a solution that is both accurate and robust, and can be computed at a rate that aligns with the needs of a control system that might make decisions based on it. The Iterative Closest Point (ICP) algorithm is widely used for this purpose, but it is susceptible to failure in practical scenarios. We present a robust and efficient solution for pose-from-point cloud estimation called the Pose Lookup Method (PLuM). PLuM is a probabilistic reward-based objective function that is resilient to measurement uncertainty and clutter. Efficiency is achieved through the use of lookup tables, which substitute complex geometric operations such as raycasting used in earlier solutions. Our results show millimetre accuracy and fast pose estimation in benchmark tests using triangulated geometry models, outperforming state-of-the-art ICP-based methods. These results are extended to field robotics applications, resulting in real-time haul truck pose estimation. By utilising point clouds from a LiDAR fixed to a rope shovel, the PLuM algorithm tracks a haul truck effectively throughout the excavation load cycle at a rate of 20 Hz, matching the sensor frame rate. PLuM is straightforward to implement and provides dependable and timely solutions in demanding environments. Full article
(This article belongs to the Special Issue Sensor Based Perception for Field Robotics)
Show Figures

Figure 1

12 pages, 2259 KB  
Article
Anhedonia in Relation to Reward and Effort Learning in Young People with Depression Symptoms
by Anna-Lena Frey, M. Siyabend Kaya, Irina Adeniyi and Ciara McCabe
Brain Sci. 2023, 13(2), 341; https://doi.org/10.3390/brainsci13020341 - 17 Feb 2023
Cited by 10 | Viewed by 7213
Abstract
Anhedonia, a central depression symptom, is associated with impairments in reward processing. However, it is not well understood which sub-components of reward processing (anticipation, motivation, consummation, and learning) are impaired in association with anhedonia in depression. In particular, it is unclear how learning [...] Read more.
Anhedonia, a central depression symptom, is associated with impairments in reward processing. However, it is not well understood which sub-components of reward processing (anticipation, motivation, consummation, and learning) are impaired in association with anhedonia in depression. In particular, it is unclear how learning about different rewards and the effort needed to obtain them might be associated with anhedonia and depression symptoms. Therefore, we examined learning in young people (N = 132, mean age 20, range 17–25 yrs.) with a range of depression and anhedonia symptoms using a probabilistic instrumental learning task. The task required participants to learn which options to choose to maximize their reward outcomes across three conditions (chocolate taste, puppy images, or money) and to minimize the physical effort required to obtain the rewards. Additionally, we collected questionnaire measures of anticipatory and consummatory anhedonia, as well as subjective reports of “liking”, “wanting” and “willingness to exert effort” for the rewards used in the task. We found that as anticipatory anhedonia increased, subjective liking and wanting of rewards decreased. Moreover, higher anticipatory anhedonia was significantly associated with lower reward learning accuracy, and participants demonstrated significantly higher reward learning than effort learning accuracy. To our knowledge, this is the first study observing an association of anhedonia with reward liking, wanting, and learning when reward and effort learning are measured simultaneously. Our findings suggest an impaired ability to learn from rewarding outcomes could contribute to anhedonia in young people. Future longitudinal research is needed to confirm this and reveal the specific aspects of reward learning that predict anhedonia. These aspects could then be targeted by novel anhedonia interventions. Full article
(This article belongs to the Special Issue Reward Processing in Health and Disease)
Show Figures

Figure 1

Back to TopTop