lnu.sePublications
Change search
Link to record
Permanent link

Direct link
Publications (10 of 69) Show all publications
Taj, S., Daudpota, S. M., Imran, A. S. & Kastrati, Z. (2025). Aspect-based sentiment analysis for software requirements elicitation using fine-tuned Bidirectional Encoder Representations from Transformers and Explainable Artificial Intelligence. Engineering applications of artificial intelligence, 151, Article ID 110632.
Open this publication in new window or tab >>Aspect-based sentiment analysis for software requirements elicitation using fine-tuned Bidirectional Encoder Representations from Transformers and Explainable Artificial Intelligence
2025 (English)In: Engineering applications of artificial intelligence, ISSN 0952-1976, E-ISSN 1873-6769, Vol. 151, article id 110632Article in journal (Refereed) Published
Abstract [en]

Aspect-Based Sentiment Analysis (ABSA) of app reviews allows a better understanding of user preferences regarding specific product features and helps the development team elicit requirements effectively. The existing literature faces challenges such as limited focus on the automation of Requirement Elicitation (RE), insufficient task-specific fine-tuning of models such as Bidirectional Encoder Representations from Transformers (BERT), and lack of interpretability owing to the black-box nature of these models. Therefore, our work makes the following significant contributions to address these challenges: (1) development and evaluation of a robust method based on ABSA for the automation of the RE process; (2) optimization of ABSA using BERT fine-tuning for enhanced performance, which includes conducting a comprehensive ablation study to obtain the best hyperparameters that guarantee the best model performance and robustness; and (3) integration of Explainable Artificial Intelligence (XAI) techniques for enhanced BERT model interpretability. Our work was evaluated on the ABSA Warehouse of Apps REviews (AWARE) dataset, a specifically tailored dataset for the RE process. Our study outperformed baseline models such as the Support Vector Machine (SVM), Convolutional Neural Network (CNN), and BERT, and achieved an average F1-Score of 0.83 for the Aspect Category Detection (ACD) task and 0.94 for the Aspect Category Polarity (ACP) task. In addition, we employed XAI using Locally Interpretable Model-Agnostic Explanations (LIME) to explain the BERT model prediction results, which aids in the improved visualization and interpretability of the app review analysis for the automated RE process.

Place, publisher, year, edition, pages
Elsevier, 2025
Keywords
Software requirement elicitation, Sentiment analysis, Aspect Category Detection, Aspect Category Polarity, App reviews, Fine-tuned Bidirectional Encoder Representations from Transformers, Explainable Artificial Intelligence
National Category
Natural Language Processing
Research subject
Computer and Information Sciences Computer Science, Information Systems
Identifiers
urn:nbn:se:lnu:diva-137445 (URN)10.1016/j.engappai.2025.110632 (DOI)001459509700001 ()2-s2.0-105001036531 (Scopus ID)
Available from: 2025-03-27 Created: 2025-03-27 Last updated: 2025-04-15Bibliographically approved
Fahad, M., Mobeen, N. E., Imran, A. S., Daudpota, S. M., Kastrati, Z., Cheikh, F. A. & Ullah, M. (2025). Deep insights into gastrointestinal health: A comprehensive analysis of GastroVision dataset using convolutional neural networks and explainable AI. Biomedical Signal Processing and Control, 102, Article ID 107260.
Open this publication in new window or tab >>Deep insights into gastrointestinal health: A comprehensive analysis of GastroVision dataset using convolutional neural networks and explainable AI
Show others...
2025 (English)In: Biomedical Signal Processing and Control, ISSN 1746-8094, E-ISSN 1746-8108, Vol. 102, article id 107260Article in journal (Refereed) Published
Abstract [en]

The gastrointestinal (GI) tract is critical in digestion and nutrient absorption, thus vital for human health. However, it is prone to diseases like cancer. Manual assessments introduce accuracy variations, consistency issues, and delays. Resources like the GastroVision dataset were introduced to advance AI in this field. Yet, it faces class imbalance issues, and baseline evaluation lacks novel methodologies, impacting accuracy. We propose a novel deep-learning model to enhance accuracy and robustness. Our approach involves averaging weights of multiple models fine-tuned with diverse hyper-parameters. In contrast to classical ensembles, our approach uses DenseNet-121 as a baseline and enables the averaging of numerous models without incurring extra inference or memory costs. Data augmentation techniques are incorporated to address class imbalance. We achieve promising results on standard performance metrics, substantially improving over baseline, notably 2.4% points in Macro Precision. Additionally, we integrate explainable AI (XAI) techniques to enhance reliability and interpretability, shedding light on the model's decision-making processes. Our study contributes to robust methodologies for imbalanced datasets, promoting model transparency and trust in predictive outcomes for clinical decision support systems.

Place, publisher, year, edition, pages
Elsevier, 2025
Keywords
GastroVision, Deep learning, Model soups, Model explainability, Generative AI, Explainable AI (XAI), Generative adversarial network (GAN), Convolutional neural network (CNN)
National Category
Biomedical Laboratory Science/Technology
Research subject
Health and Caring Sciences, Health Informatics
Identifiers
urn:nbn:se:lnu:diva-133748 (URN)10.1016/j.bspc.2024.107260 (DOI)001373625000001 ()2-s2.0-85211021231 (Scopus ID)
Available from: 2024-12-04 Created: 2024-12-04 Last updated: 2024-12-20Bibliographically approved
Kastrati, M., Imran, A. S., Hashmi, E., Kastrati, Z., Daudpota, S. M. & Biba, M. (2025). Unlocking language barriers: Assessing pre-trained large language models across multilingual tasks and unveiling the black box with Explainable Artificial Intelligence. Engineering applications of artificial intelligence, 149, Article ID 110136.
Open this publication in new window or tab >>Unlocking language barriers: Assessing pre-trained large language models across multilingual tasks and unveiling the black box with Explainable Artificial Intelligence
Show others...
2025 (English)In: Engineering applications of artificial intelligence, ISSN 0952-1976, E-ISSN 1873-6769, Vol. 149, article id 110136Article in journal (Refereed) Published
Abstract [en]

Large Language Models (LLMs) have revolutionized many industrial applications and paved the way for fostering a new research direction in many fields. Conventional Natural Language Processing (NLP) techniques, for instance, are no longer necessary for many text-based tasks, including polarity estimation, sentiment and emotion classification, and hate speech detection. However, training a language model for domain-specific tasks is hugely costly and requires high computational power, thereby restricting its true potential for standard tasks. This study, therefore, provides a comprehensive analysis of the latest pre-trained LLMs for various NLP-related applications without fine-tuning them to evaluate their effectiveness. Five language models are thus employed in this study on six distinct NLP tasks (including emotion recognition, sentiment analysis, hate speech detection, irony detection, offensiveness detection, and stance detection) for 12 languages from low- to medium- and high-resource. Generative Pre-trained Transformer 4 (GPT-4) and Gemini Pro outperform state-of-the-art models, achieving average F1 scores of 70.6% and 68.8% on the Tweet Sentiment Multilingual dataset compared to the state-of-the-art average F1 score of 66.8%. The study further interprets the findings obtained by the LLMs using Explainable Artificial Intelligence (XAI). To the best of our knowledge, it is the first time any study has employed explainability on pre-trained language models.

Place, publisher, year, edition, pages
Elsevier, 2025
Keywords
Large language models, Zero-shot classification, Explainable Artificial Intelligence, Sentiment analysis, Emotion recognition
National Category
Natural Language Processing
Research subject
Computer and Information Sciences Computer Science, Information Systems
Identifiers
urn:nbn:se:lnu:diva-137213 (URN)10.1016/j.engappai.2025.110136 (DOI)001446400200001 ()2-s2.0-86000570396 (Scopus ID)
Available from: 2025-03-12 Created: 2025-03-12 Last updated: 2025-03-27Bibliographically approved
Kastrati, Z., Fatima, S., Kurti, A., Daudpota, S. M. & Imran, A. S. (2024). Analyzing and Predicting the Helpfulness of Reviews in MOOCs Context Using Deep Learning. In: Jonathan Flearmoy (Ed.), Procedia Computer Science 246: 28th International Conference on Knowledge Based and Intelligent information and Engineering Systems (KES 2024). Paper presented at 28th International Conference on Knowledge Based and Intelligent information and Engineering Systems (KES 2024) (pp. 772-781). Elsevier, 246
Open this publication in new window or tab >>Analyzing and Predicting the Helpfulness of Reviews in MOOCs Context Using Deep Learning
Show others...
2024 (English)In: Procedia Computer Science 246: 28th International Conference on Knowledge Based and Intelligent information and Engineering Systems (KES 2024) / [ed] Jonathan Flearmoy, Elsevier, 2024, Vol. 246, p. 772-781Conference paper, Published paper (Refereed)
Abstract [en]

Students' feedback is an essential part of the teaching-learning process and serves as an effective instrument for continuous improvement in educational environments. The insights gathered from students' experiences and perceptions expressed in reviews provide instructors with a valuable resource to enhance their teaching methods, instructional design, and overall classroom dynamics. However, students' reviews are often unclear, contradictory, and conflicting with each other, making their interpretation and use challenging. Therefore, this study proposes a novel deep learning-based approach that helps course designers and instructors effectively identify constructive and useful reviews. The approach leverages the integration of several attributes, including textual review, student satisfaction, meta-data of the course, and review-derived information such as sentiment, readability, and review depth. The approach is tested on a real-life dataset comprising 38,717 reviews gathered from the Coursera learning platform for the purpose of this study. The experimental results, with an F1-score of 0.91, suggest that the approach can be an effective tool for educators and instructional designers to identify helpful student reviews.

Place, publisher, year, edition, pages
Elsevier, 2024
Series
Procedia Computer Science, E-ISSN 1877-0509 ; 246
Keywords
Review helpfulness, MOOCs, Deep learning, Student satisfaction, Student's feedback, Review depth, Sentiment, Meta-data
National Category
Information Systems
Research subject
Computer and Information Sciences Computer Science, Computer Science
Identifiers
urn:nbn:se:lnu:diva-133565 (URN)10.1016/j.procs.2024.09.496 (DOI)2-s2.0-85213356696 (Scopus ID)
Conference
28th International Conference on Knowledge Based and Intelligent information and Engineering Systems (KES 2024)
Available from: 2024-11-29 Created: 2024-11-29 Last updated: 2025-02-27Bibliographically approved
Akhlaq, F., Ali, S., Imran, A. S., Daudpota, S. M. & Kastrati, Z. (2024). Diving Deep into Bone Anomalies on the FracAtlas Dataset Using Deep Learning and Explainable AI. In: Proceedings of the 2024 International Conference on Engineering & Computing Technologies (ICECT): . Paper presented at 2024 International Conference on Engineering & Computing Technologies (ICECT) (pp. 1-6). Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>Diving Deep into Bone Anomalies on the FracAtlas Dataset Using Deep Learning and Explainable AI
Show others...
2024 (English)In: Proceedings of the 2024 International Conference on Engineering & Computing Technologies (ICECT), Institute of Electrical and Electronics Engineers (IEEE), 2024, p. 1-6Conference paper, Published paper (Refereed)
Abstract [en]

Medical image analysis has undergone significant advancements with the integration of machine learning techniques, particularly in the realm of bone anomaly detection. The availability of recent datasets and the lack of benchmarking and explainability components provide numerous opportunities in this domain. This study proposes a benchmarking approach to a recently published FracAtlas dataset utilizing state-of-the-art deep-learning models coupled with explainable artificial intelligence (XAI) having two distinct modules. The first module involves the binary classification of fractures in different body parts and explains the decision-making process of the best-performing model using an XAI technique known as EigenCAM. EigenCAM generates heatmaps on every layer of the YOLOv8m model to explain how the model reached a conclusion and localizes the fracture using a heatmap. To verify the heatmap, we also detected fractures using the YOLOv8m detection model, which achieved a mAP@O.5 of 59.5%, outperforming the baseline results on this dataset. The second module involves a multi-class classification task to categorize images into one of the five anatomical regions. The best-performing model for binary classification is the YOLOv8m model, with an accuracy of 83.1%, whereas the best-performing model for multi-class classification is the YOLOv8s, achieving an accuracy of 96.2%.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2024
Keywords
Heating systems, Deep learning, Accuracy, Image analysis, Explainable AI, Computational modeling, Decision making, Fracture classification and detection, Medical imaging, X-rays, Explainable AI, Deep learning
National Category
Medical Imaging
Identifiers
urn:nbn:se:lnu:diva-131439 (URN)10.1109/ICECT61618.2024.10581288 (DOI)2-s2.0-85199160052 (Scopus ID)
Conference
2024 International Conference on Engineering & Computing Technologies (ICECT)
Available from: 2024-07-14 Created: 2024-07-14 Last updated: 2025-02-12Bibliographically approved
Ahmed, A., Imran, A. S., Manaf, A., Kastrati, Z. & Daudpota, S. M. (2024). Enhancing wrist abnormality detection with YOLO: Analysis of state-of-the-art single-stage detection models. Biomedical Signal Processing and Control, 93, Article ID 106144.
Open this publication in new window or tab >>Enhancing wrist abnormality detection with YOLO: Analysis of state-of-the-art single-stage detection models
Show others...
2024 (English)In: Biomedical Signal Processing and Control, ISSN 1746-8094, E-ISSN 1746-8108, Vol. 93, article id 106144Article in journal (Refereed) Published
Abstract [en]

Diagnosing and treating abnormalities in the wrist, specifically distal radius, and ulna fractures, is a crucial concern among children, adolescents, and young adults, with a higher incidence rate during puberty. However, the scarcity of radiologists and the lack of specialized training among medical professionals pose a significant risk to patient care. This problem is further exacerbated by the rising number of imaging studies and limited access to specialist reporting in certain regions. This highlights the need for innovative solutions to improve the diagnosis and treatment of wrist abnormalities. Automated wrist fracture detection using object detection has shown potential, but current studies mainly use two-stage detection methods with limited evidence for single-stage effectiveness. This study employs state-of-the-art single-stage deep neural network-based detection models YOLOv5, YOLOv6, YOLOv7, and YOLOv8 to detect wrist abnormalities. Through extensive experimentation, we found that these YOLO models outperform the commonly used two-stage detection algorithm, Faster R-CNN, in fracture detection. Additionally, compound-scaled variants of each YOLO model were compared, with YOLOv8 m demonstrating a highest fracture detection sensitivity of 0.92 and mean average precision (mAP) of 0.95. On the other hand, YOLOv6 m achieved the highest sensitivity across all classes at 0.83. Meanwhile, YOLOv8x recorded the highest mAP of 0.77 for all classes on the GRAZPEDWRI-DX pediatric wrist dataset, highlighting the potential of single-stage models for enhancing pediatric wrist imaging.

Place, publisher, year, edition, pages
Elsevier, 2024
Keywords
Wrist fracture detection, Object localization, Medical imaging, Pediatric X-ray, Deep learning, YOLO
National Category
Information Systems
Research subject
Computer Science, Information and software visualization
Identifiers
urn:nbn:se:lnu:diva-128002 (URN)10.1016/j.bspc.2024.106144 (DOI)001192255500001 ()2-s2.0-85185833569 (Scopus ID)
Available from: 2024-02-23 Created: 2024-02-23 Last updated: 2025-02-12Bibliographically approved
Ali, S., Imran, A. S., Kastrati, Z., Daudpota, S. M., Cheikh, F. A. & Ullah, M. (2024). Enhancing Wrist Fracture Detection and Classification through Deep Learning and XAI. In: 2024 12th European Workshop on Visual Information Processing (EUVIP): . Paper presented at 12th European Workshop on Visual Information Processing (EUVIP). IEEE
Open this publication in new window or tab >>Enhancing Wrist Fracture Detection and Classification through Deep Learning and XAI
Show others...
2024 (English)In: 2024 12th European Workshop on Visual Information Processing (EUVIP), IEEE, 2024Conference paper, Published paper (Refereed)
Abstract [en]

According to WHO, approximately 1.71 billion people worldwide have musculoskeletal conditions, which include various issues such as intact muscles, bones, joints, and fractures. Among those, fractures are the most common. Due to the emergency nature of diagnosing fractures, there is a high chance of misdiagnosis for several reasons, such as the unavailability of radiologists, physicians' lack of experience, and other factors. Fracture diagnostic or X-ray interpretation errors can be reduced if radiographs are always read instantly by radiologists or automatically. In our study, we are focusing on automating wrist fracture diagnosis, where we are utilizing the publicly available GRAZPEDWRI-DX dataset, which consists of 20,327 wrist radiographs. We employed the YOLOv9 model for fracture detection, achieving an mAP@50 of 0.677, which surpasses previous benchmarks. For fracture classification, we trained several state-of-the-art deep learning models, including VGG16, VGG19, ResNet50, EfficientNetB7, DenseNet121, MobileNet, and ConvNeXtXLarge. In particular, the YOLOv8-cls model surpassed all others in accuracy (0.93), precision (0.9352), and recall (0.8855), reaching its peak performance at epoch 70. To elucidate the decision-making process of both the detection and classification models, we generated explainable saliency maps through EigenCAM, making the models more explainable and interpretable.

Place, publisher, year, edition, pages
IEEE, 2024
Keywords
wrist fracture detection, wrist fracture classification, deep learning, computer vision, explainable artificial intelligence
National Category
Orthopaedics Computer Sciences
Identifiers
urn:nbn:se:lnu:diva-133897 (URN)10.1109/EUVIP61797.2024.10772888 (DOI)2-s2.0-85214661671 (Scopus ID)
Conference
12th European Workshop on Visual Information Processing (EUVIP)
Available from: 2024-12-11 Created: 2024-12-11 Last updated: 2025-02-12Bibliographically approved
Ahmed, A., Imran, A. S., Kastrati, Z., Daudpota, S. M., Ullah, M. & Noor, W. (2024). Learning from the few: Fine-grained approach to pediatric wrist pathology recognition on a limited dataset. Computers in Biology and Medicine, 181, Article ID 109044.
Open this publication in new window or tab >>Learning from the few: Fine-grained approach to pediatric wrist pathology recognition on a limited dataset
Show others...
2024 (English)In: Computers in Biology and Medicine, ISSN 0010-4825, E-ISSN 1879-0534, Vol. 181, article id 109044Article in journal (Refereed) Published
Abstract [en]

Wrist pathologies, particularly fractures common among children and adolescents, present a critical diagnostic challenge. While X-ray imaging remains a prevalent diagnostic tool, the increasing misinterpretation rates highlight the need for more accurate analysis, especially considering the lack of specialized training among many surgeons and physicians. Recent advancements in deep convolutional neural networks offer promise in automating pathology detection in trauma X-rays. However, distinguishing subtle variations between pediatric wrist pathologies in X-rays remains challenging. Traditional manual annotation, though effective, is laborious, costly, and requires specialized expertise. In this paper, we address the challenge of pediatric wrist pathology recognition with a fine-grained approach, aimed at automatically identifying discriminative regions in X-rays without manual intervention. We refine our fine-grained architecture through ablation analysis and the integration of LION. Leveraging Grad-CAM, an explainable AI technique, we highlight these regions. Despite using limited data, reflective of real-world medical study constraints, our method consistently outperforms state-of-the-art image recognition models on both augmented and original (challenging) test sets. Our proposed refined architecture achieves an increase in accuracy of 1.06% and 1.25% compared to the baseline method, resulting in accuracies of 86% and 84%, respectively. Moreover, our approach demonstrates the highest fracture sensitivity of 97%, highlighting its potential to enhance wrist pathology recognition.

Place, publisher, year, edition, pages
Elsevier, 2024
Keywords
Fine-grained visual classification, Wrist X-ray imaging, Explainable artificial intelligence (XAI), Fracture recognition, Medical, Deep learning
National Category
Medical Imaging
Research subject
Health and Caring Sciences, Health Informatics
Identifiers
urn:nbn:se:lnu:diva-132090 (URN)10.1016/j.compbiomed.2024.109044 (DOI)2-s2.0-85202149482 (Scopus ID)
Available from: 2024-08-24 Created: 2024-08-24 Last updated: 2025-02-12Bibliographically approved
Kastrati, M., Kastrati, Z., Shariq Imran, A. & Biba, M. (2024). Leveraging distant supervision and deep learning for twitter sentiment and emotion classification. Journal of Intelligent Information Systems, 62, 1045-1070
Open this publication in new window or tab >>Leveraging distant supervision and deep learning for twitter sentiment and emotion classification
2024 (English)In: Journal of Intelligent Information Systems, ISSN 0925-9902, E-ISSN 1573-7675, Vol. 62, p. 1045-1070Article in journal (Refereed) Published
Abstract [en]

Nowadays, various applications across industries, healthcare, and security have begun adopting automatic sentiment analysis and emotion detection in short texts, such as posts from social media. Twitter stands out as one of the most popular online social media platforms due to its easy, unique, and advanced accessibility using the API. On the other hand, supervised learning is the most widely used paradigm for tasks involving sentiment polarity and fine-grained emotion detection in short and informal texts, such as Twitter posts. However, supervised learning models are data-hungry and heavily reliant on abundant labeled data, which remains a challenge. This study aims to address this challenge by creating a large-scale real-world dataset of 17.5 million tweets. A distant supervision approach relying on emojis available in tweets is applied to label tweets corresponding to Ekman’s six basic emotions. Additionally, we conducted a series of experiments using various conventional machine learning models and deep learning, including transformer-based models, on our dataset to establish baseline results. The experimental results and an extensive ablation analysis on the dataset showed that BiLSTM with FastText and an attention mechanism outperforms other models in both classification tasks, achieving an F1-score of 70.92% for sentiment classification and 54.85% for emotion detection.

Place, publisher, year, edition, pages
Springer, 2024
Keywords
Distant supervision, Emotion detection, Sentiment analysis, Deep learning, Transformers, Twitter, Emojis
National Category
Information Systems
Research subject
Computer and Information Sciences Computer Science, Information Systems
Identifiers
urn:nbn:se:lnu:diva-128396 (URN)10.1007/s10844-024-00845-0 (DOI)001190186000001 ()2-s2.0-85188512613 (Scopus ID)
Available from: 2024-03-22 Created: 2024-03-22 Last updated: 2025-02-12Bibliographically approved
Ahmed, A., Imran, A. S., Ullah, M., Kastrati, Z. & Daudpota, S. M. (2024). Navigating Limitations With Precision: A Fine-Grained Ensemble Approach To Wrist Pathology Recognition On A Limited X-Ray Dataset. In: 2024 IEEE International Conference on Image Processing (ICIP): . Paper presented at 2024 IEEE International Conference on Image Processing (ICIP) (pp. 3077-3083). IEEE
Open this publication in new window or tab >>Navigating Limitations With Precision: A Fine-Grained Ensemble Approach To Wrist Pathology Recognition On A Limited X-Ray Dataset
Show others...
2024 (English)In: 2024 IEEE International Conference on Image Processing (ICIP), IEEE, 2024, p. 3077-3083Conference paper, Published paper (Refereed)
Abstract [en]

The exploration of automated wrist fracture recognition has gained considerable research attention in recent years. In practical medical scenarios, physicians and surgeons may lack the specialized expertise required for accurate X-ray interpretation, highlighting the need for machine vision to enhance diagnostic accuracy. However, conventional recognition techniques face challenges in discerning subtle differences in X-rays when classifying wrist pathologies, as many of these pathologies, such as fractures, can be small and hard to distinguish. This study tackles wrist pathology recognition as a fine-grained visual recognition (FGVR) problem, utilizing a limited, custom-curated dataset that mirrors real-world medical constraints, relying solely on image-level annotations. We introduce a specialized FGVR-based ensemble approach to identify discriminative regions within X-rays. We employ an Explainable AI (XAI) technique called Grad-CAM to pinpoint these regions. Our ensemble approach outperformed many conventional SOTA and FGVR techniques, underscoring the effectiveness of our strategy in enhancing accuracy in wrist pathology recognition.

Place, publisher, year, edition, pages
IEEE, 2024
Keywords
Wrist, Pathology, Visualization, Image recognition, Accuracy, Explainable AI, Face recognition, Fine-grained visual classification, Medical xray imaging, Explainable artificial intelligence (XAI), Fracture recognition, Deep ensemble learning
National Category
Bioinformatics (Computational Biology)
Research subject
Computer and Information Sciences Computer Science, Information Systems; Natural Science, Biomedical Sciences
Identifiers
urn:nbn:se:lnu:diva-133200 (URN)10.1109/ICIP51287.2024.10648070 (DOI)2-s2.0-85216871535 (Scopus ID)
Conference
2024 IEEE International Conference on Image Processing (ICIP)
Available from: 2024-10-31 Created: 2024-10-31 Last updated: 2025-03-05Bibliographically approved
Organisations
Identifiers
ORCID iD: ORCID iD iconorcid.org/0000-0002-0199-2377

Search in DiVA

Show all publications