Connectivity Fund

Improving Cross-Lingual Retrieval of Previously Fact-Checked Claims

Róbert Móro Researcher at Kempelen Institute of Intelligent Technologies To mitigate disinformation with AI in a trustworthy way, it should prioritize human agency and control, transparency, and accountability including the means for redress. This can be achieved by using AI to support rather than to replace media professionals, such as fact-checkers, in their efforts to […]

Improving Cross-Lingual Retrieval of Previously Fact-Checked Claims Read More »

Enhancing Reliability and Trustworthiness in IoT Applications through Deep Learning-Based Data Imputation Techniques

Hakob Grigoryan PhD at NVISION With the evolution of intelligent sensing devices and the Internet of Things (IoT), a vast amount of data is generated from various sources, including sensors, cameras, and network infrastructures, and is transmitted to servers for analysis. Data streaming from sensors in IoT systems might face quality issues like incompleteness due

Enhancing Reliability and Trustworthiness in IoT Applications through Deep Learning-Based Data Imputation Techniques Read More »

Exploring Intrusion Detection Knowledge Transfer Between Network Environments

Patrik Goldschmidt PhD candidate at Kempelen Institute of Intelligent Technologies With the rise of information technology and the Internet, the number of cybersecurity incidents has grown immensely. As a response, the research area of Intrusion Detection Systems (IDSs), aiming to detect and mitigate cyber threats, has gained significant attention. Our research focuses on Network IDSs

Exploring Intrusion Detection Knowledge Transfer Between Network Environments Read More »

Enhancing Trustworthiness in Healthcare Large Language Models

Muhammad Waseem Postdoctoral Researcher at Faculty of Information Technology, University of Jyväskylä, Jyväskylä, Finland Large Language Models (LLMs) are advanced AI tools capable of understanding and generating human-like text, advancing various sectors, including healthcare. This project aims to enhance healthcare services using LLMs, focusing on improving their trustworthiness for clinical applications. Trustworthiness encompasses reliability, fairness,

Enhancing Trustworthiness in Healthcare Large Language Models Read More »

Continual Self-Supervised Learning

Giacomo Cignoni Research Fellow at the University of Pisa Learning continually from non-stationary data streams is a challenging research topic of growing popularity in the last few years. Being able to learn, adapt and generalize continually, in an efficient way appears to be fundamental for a more sustainable development of Artificial Intelligent systems. However, research

Continual Self-Supervised Learning Read More »

Tractable and Explainable Probabilistic AI

Lennert De Smet PhD at KU Leuven Transparency and technical robustness are two fundamental requirements for AI systems following the European Union AI Act, especially in higher-risk domains. Transparency is intricately related to the notion of explainability, allowing an AI system to accurately describe the reasoning behind its predictions. Through such explanations does the system

Tractable and Explainable Probabilistic AI Read More »

Trustworthy, Ethical and Beneficial-to-All Multiagent Systems Solutions for Social Ridesharing and the Hospitality Industry

Georgios Chalkiadakis Professor at Technical University of Crete Current mobility-as-a-service platforms have departed from the original objectives of the sharing economy-inspired social ridesharing paradigm: regrettably, they view drivers as taxi workers; focus on profit maximization rather than fair travel costs’ allocation; and disregard essential private preferences of users (relating for instance to their feeling of

Trustworthy, Ethical and Beneficial-to-All Multiagent Systems Solutions for Social Ridesharing and the Hospitality Industry Read More »

Translating between AI Evaluation and Job Tasks in the human workplace for trustworthy and reliable AI deployment

Marko Tesic Post-doc at LCFI, University of Cambridge, UK Recent advancements in AI, particularly in language modeling, have rekindled concerns about the potential automation of certain roles within the human workforce. To better understand which roles are susceptible to automation and to ensure the trustworthy and reliable deployment of AI, I aim to establish a

Translating between AI Evaluation and Job Tasks in the human workplace for trustworthy and reliable AI deployment Read More »

Evaluation of cognitive capabilities for LLMs

Lorenzo Pacchiardi Post-doc at University of Cambridge Artificial Intelligence (AI) systems (such as reinforcement-learning agents and Large Language Models, or LLMs) are typically evaluated by testing them on a benchmark and reporting an aggregated score. As benchmarks are constituted of instances demanding various capability levels to be completed, the aggregated score is uninformative of the

Evaluation of cognitive capabilities for LLMs Read More »