We foresee robots that bootstrap knowledge representations and use them for classifying relevant situations and making decisions based on future observations. Particularly for assistive robots, the bootstrapping mechanism might be supervised by humans who should not repeat a training phase several times and should be able to refine the taught representation. We consider robots that bootstrap structured representations to classify some intelligible categories. Such a structure should be incrementally bootstrapped, i.e., without invalidating the identified category models when a new additional category is considered. To tackle this scenario, we presented the Scene Identification and Tagging (SIT) algorithm, which bootstraps structured knowledge representation in a crisp OWL-DL ontology. Over time, SIT bootstraps a graph representing scenes, sub-scenes and similar scenes. Then, SIT can classify new scenes within the bootstrapped graph through logic-based reasoning. However, SIT has issues with sensory data because its crisp implementation is not robust to perception noises. This paper presents a reformulation of SIT within the fuzzy domain, which exploits a fuzzy DL ontology to overcome the robustness issues. By comparing the performances of fuzzy and crisp implementations of SIT, we show that fuzzy SIT is robust, preserves the properties of its crisp formulation, and enhances the bootstrapped representations. On the contrary, the fuzzy implementation of SIT leads to less intelligible knowledge representations than the one bootstrapped in the crisp domain.
我们预计将出现能够引导知识表示的机器人,并将其用于分类相关情况并根据未来观察结果做出决策的机器人。特别是辅助机器人,引导机制可能由人类监督,他们不应该重复训练阶段多次,并且应该能够精炼所教授的表示。我们认为,引导结构化表示以分类一些可解释类别的机器人。这种结构应该通过逐步引导来进行,即在考虑新增类别时不会破坏已确定的类别模型。为解决这种情况,我们提出了Scene Identification and Tagging (SIT)算法,它在 crisp OWL-DL 上下文中引导结构化知识表示。随着时间的推移,SIT 通过基于逻辑推理绘制场景、子场景和类似场景的图。然后,SIT 通过逻辑推理对引导的图中的新场景进行分类。然而,SIT 在感官数据方面存在问题,因为其明确的实现对感知噪声不具有鲁棒性。本文在模糊领域对SIT进行了重新表述,利用模糊DL 上下文克服了鲁棒性问题。通过比较模糊和明确实现SIT的性能,我们证明了模糊SIT具有鲁棒性,保留了其明确的公式的性质,并增强了引导的表示。相反,模糊实现SIT导致生成的知识表示比在清晰领域引导的更不清晰。
https://arxiv.org/abs/2404.11744
Ontology alignment, a critical process in the Semantic Web for detecting relationships between different ontologies, has traditionally focused on identifying so-called "simple" 1-to-1 relationships through class labels and properties comparison. The more practically useful exploration of more complex alignments remains a hard problem to automate, and as such is largely underexplored, i.e. in application practice it is usually done manually by ontology and domain experts. Recently, the surge in Natural Language Processing (NLP) capabilities, driven by advancements in Large Language Models (LLMs), presents new opportunities for enhancing ontology engineering practices, including ontology alignment tasks. This paper investigates the application of LLM technologies to tackle the complex ontology alignment challenge. Leveraging a prompt-based approach and integrating rich ontology content so-called modules our work constitutes a significant advance towards automating the complex alignment task.
知识图谱对齐,作为一个在语义网中检测不同知识图谱之间关系的关键过程,通常集中在通过类标签和属性比较识别所谓的“简单”1对1关系。更实际可行的对更复杂对齐的探索仍然是一个难以自动化的困难问题,因此它仍然被大大忽视。即在应用实践中,通常是由本体和领域专家手动完成的。最近,自然语言处理(NLP)能力的突飞猛进,受到大型语言模型(LLMs)的进步,为增强语义工程实践提供了新的机会,包括语义对齐任务。本文调查了LLM技术在解决复杂语义对齐挑战中的应用。我们利用提示式方法并整合了丰富的语义内容,所谓的模块,这使得我们的工作在自动解决复杂对齐任务方面取得了显著的进展。
https://arxiv.org/abs/2404.10329
Ontology Matching (OM), is a critical task in knowledge integration, where aligning heterogeneous ontologies facilitates data interoperability and knowledge sharing. Traditional OM systems often rely on expert knowledge or predictive models, with limited exploration of the potential of Large Language Models (LLMs). We present the LLMs4OM framework, a novel approach to evaluate the effectiveness of LLMs in OM tasks. This framework utilizes two modules for retrieval and matching, respectively, enhanced by zero-shot prompting across three ontology representations: concept, concept-parent, and concept-children. Through comprehensive evaluations using 20 OM datasets from various domains, we demonstrate that LLMs, under the LLMs4OM framework, can match and even surpass the performance of traditional OM systems, particularly in complex matching scenarios. Our results highlight the potential of LLMs to significantly contribute to the field of OM.
知识集成中的元数据匹配(OM)是一个关键任务,其中对异构知识本体的对齐有助于促进数据互操作性和知识共享。传统的OM系统通常依赖于专家知识或预测模型,对大型语言模型的潜力探索有限。我们提出了LLMs4OM框架,一种评估LLM在OM任务中有效性的新方法。该框架采用两个模块进行检索和匹配,分别通过三个知识表示层的零散提示进行加强:概念、概念父体和概念子体。通过使用各种领域的20个OM数据集进行全面评估,我们证明了LLM在LLMs4OM框架下可以匹配甚至超过传统OM系统的表现,特别是在复杂匹配场景中。我们的结果突出了LLM在OM领域显著贡献的潜力。
https://arxiv.org/abs/2404.10317
The paper tackles the issue of mapping logic axioms formalised in the Ontology Web Language (OWL) within the Object-Oriented Programming (OOP) paradigm. The issues of mapping OWL axioms hierarchies and OOP objects hierarchies are due to OWL-based reasoning algorithms, which might change an OWL hierarchy at runtime; instead, OOP hierarchies are usually defined as static structures. Although programming paradigms based on reflection allow changing the OOP hierarchies at runtime and mapping OWL axioms dynamically, there are no currently available mechanisms that do not limit the reasoning algorithms. Thus, the factory-based paradigm is typically used since it decouples the OWL and OOP hierarchies. However, the factory inhibits OOP polymorphism and introduces a paradigm shift with respect to widely accepted OOP paradigms. We present the OWLOOP API, which exploits the factory to not limit reasoning algorithms, and it provides novel OOP interfaces concerning the axioms in an ontology. OWLOOP is designed to limit the paradigm shift required for using ontologies while improving, through OOP-like polymorphism, the modularity of software architectures that exploit logic reasoning. The paper details our OWL to OOP mapping mechanism, and it shows the benefits and limitations of OWLOOP through examples concerning a robot in a smart environment.
本文研究了在面向对象编程(OOP)范式内,将语义知识图谱(OWL)中的推理规则映射到OWL模型的逻辑轴理问题。OWL轴理层次结构和OOP对象层次结构的映射问题是因为基于OWL的推理算法可能会在运行时改变OWL层次结构;而OOP层次结构通常被定义为静态结构。尽管基于反思的编程范式允许在运行时改变OOP层次结构,并动态地映射OWL轴理,但目前没有可用的机制不限制推理算法。因此,通常是基于工厂的方法,因为它解耦了OWL和OOP层次结构。然而,工厂会抑制OOP多态性,并引入与广泛接受的多范式OOP范式不同的范式转变。我们提出了OWLOOP API,该API利用工厂来避免限制推理算法,并提供了关于语义模型中轴理的新颖OOP接口。OWLOOP旨在通过类似的OOP方式限制使用语义模型的范式转变,同时提高软件架构的模块性,通过逻辑推理来利用。本文详细介绍了我们的OWL到OOP映射机制,并通过一个智能环境中的人工机器人示例,展示了OWLOOP的优势和局限性。
https://arxiv.org/abs/2404.09305
Effective ontology transfer has been a major goal of recent work on event argument extraction (EAE). Two methods in particular -- question answering (QA) and template infilling (TI) -- have emerged as promising approaches to this problem. However, detailed explorations of these techniques' ability to actually enable this transfer are lacking. In this work, we provide such a study, exploring zero-shot transfer using both techniques on six major EAE datasets at both the sentence and document levels. Further, we challenge the growing reliance on LLMs for zero-shot extraction, showing that vastly smaller models trained on an appropriate source ontology can yield zero-shot performance superior to that of GPT-3.5 or GPT-4.
有效的本体转移一直是事件论证提取(EAE)领域最近工作的主要目标。尤其是问答(QA)和模板填充(TI)两种方法——被认为是解决这个问题的有前途的方法。然而,这些技术实际实现这一转移的能力的详细探讨还缺乏。在这项工作中,我们提供了这样的研究,探讨了在句子和文档级别上使用这两种技术进行零散转移。此外,我们还挑战了越来越多地依赖LLM进行零散提取的趋势,证明了在适当的本体架构上训练的小规模模型可以产生与GPT-3.5或GPT-4.0相当甚至更好的零散性能。
https://arxiv.org/abs/2404.08579
The creation of high-quality ontologies is crucial for data integration and knowledge-based reasoning, specifically in the context of the rising data economy. However, automatic ontology matchers are often bound to the heuristics they are based on, leaving many matches unidentified. Interactive ontology matching systems involving human experts have been introduced, but they do not solve the fundamental issue of flexibly finding additional matches outside the scope of the implemented heuristics, even though this is highly demanded in industrial settings. Active machine learning methods appear to be a promising path towards a flexible interactive ontology matcher. However, off-the-shelf active learning mechanisms suffer from low query efficiency due to extreme class imbalance, resulting in a last-mile problem where high human effort is required to identify the remaining matches. To address the last-mile problem, this work introduces DualLoop, an active learning method tailored to ontology matching. DualLoop offers three main contributions: (1) an ensemble of tunable heuristic matchers, (2) a short-term learner with a novel query strategy adapted to highly imbalanced data, and (3) long-term learners to explore potential matches by creating and tuning new heuristics. We evaluated DualLoop on three datasets of varying sizes and domains. Compared to existing active learning methods, we consistently achieved better F1 scores and recall, reducing the expected query cost spent on finding 90% of all matches by over 50%. Compared to traditional interactive ontology matchers, we are able to find additional, last-mile matches. Finally, we detail the successful deployment of our approach within an actual product and report its operational performance results within the Architecture, Engineering, and Construction (AEC) industry sector, showcasing its practical value and efficiency.
高质量本体论的创建对于数据集成和基于知识的推理至关重要,尤其是在数据经济迅速崛起的背景下。然而,自动本体论匹配器通常受到其基于的启发式约束,导致许多匹配无法确定。已经引入了涉及人类专家的交互式本体论匹配系统,但这些系统并未解决实施启发式约束的基本问题,尽管在工业环境中这一点非常重要。积极机器学习方法似乎是通往具有灵活性的交互式本体论匹配器的有望之路。然而,由于极端的类别不平衡,现成的积极学习机制导致查询效率较低,导致最后1公里问题,需要高人类努力来确定剩余的匹配。为解决最后1公里问题,本文引入了DualLoop,一种专为本体论匹配的积极学习方法。DualLoop 带来了三个主要贡献:(1)可调整的启发式匹配器的集合;(2)适应高度不平衡数据的新查询策略;(3)创建并调整新本体论以探索潜在匹配。我们在三个不同规模和领域的数据集上评估了DualLoop。与现有积极学习方法相比,我们始终获得了更好的F1分数和召回,将预计查询成本用于找到90%的匹配降低了50%以上。与传统交互式本体论匹配器相比,我们能够找到额外的最后1公里匹配。最后,我们详细介绍了将我们的方法成功部署在实际产品中的情况,并报告了其在建筑、工程和 Construction(AEC)行业部门中的操作性能结果,展示了其实用价值和效率。
https://arxiv.org/abs/2404.07663
Sourcing and identification of new manufacturing partners is crucial for manufacturing system integrators to enhance agility and reduce risk through supply chain diversification in the global economy. The advent of advanced large language models has captured significant interest, due to their ability to generate comprehensive and articulate responses across a wide range of knowledge domains. However, the system often falls short in accuracy and completeness when responding to domain-specific inquiries, particularly in areas like manufacturing service discovery. This research explores the potential of leveraging Knowledge Graphs in conjunction with ChatGPT to streamline the process for prospective clients in identifying small manufacturing enterprises. In this study, we propose a method that integrates bottom-up ontology with advanced machine learning models to develop a Manufacturing Service Knowledge Graph from an array of structured and unstructured data sources, including the digital footprints of small-scale manufacturers throughout North America. The Knowledge Graph and the learned graph embedding vectors are leveraged to tackle intricate queries within the digital supply chain network, responding with enhanced reliability and greater interpretability. The approach highlighted is scalable to millions of entities that can be distributed to form a global Manufacturing Service Knowledge Network Graph that can potentially interconnect multiple types of Knowledge Graphs that span industry sectors, geopolitical boundaries, and business domains. The dataset developed for this study, now publicly accessible, encompasses more than 13,000 manufacturers' weblinks, manufacturing services, certifications, and location entity types.
采购和识别新制造商合作伙伴对全球经济中的供应链多元化至关重要,这可以提高制造系统集成商的敏捷性,并通过供应链多元化提高风险降低。先进的大型语言模型的出现引起了广泛关注,因为它们能够生成全面且明确的回答,涵盖广泛的领域知识。然而,当回答领域特定问题时,系统往往存在准确性和完整性不足的情况,特别是在制造业服务发现领域。这项研究探讨了在知识图谱与 ChatGPT 的结合下,简化潜在客户在识别小制造企业过程中的可能性。 在本研究中,我们提出了一种方法,将自下而上的本体与先进机器学习模型相结合,从包括北美地区小型制造商的数字足迹在内的一系列结构和非结构化数据源中开发出制造业服务知识图。知识图和学到的图嵌入向量被用来处理数字供应链网络中的复杂查询,并回应提高可靠性和增强可解释性的答案。 所提出的方法具有可扩展性,可以将数百万实体分配到形成一个全球制造业服务知识网络图,这个网络图可能连接多个跨越行业部门、地理政治边界和企业领域的知识图。 为这项研究创建的数据集,现已成为公开可访问的数据库,包括13,000多个制造商网站、制造业服务、认证和位置实体类型。
https://arxiv.org/abs/2404.06571
In this paper we present a publicly-available maintenance ontology (Iof-maint). Iof-maint is a modular ontology aligned with the Industrial Ontology Foundry Core (IOF Core) and contains 20 classes and 2 relations. It provides a set of maintenance-specific terms used in a wide variety of practical data-driven use cases. Iof-maint supports OWL DL reasoning, is documented, and is actively maintained on GitHub. In this paper, we describe the evolution of the Iof-maint reference ontology based on the extraction of common concepts identified in a number of application ontologies working with industry maintenance work order, procedure and failure mode data.
在本文中,我们提出了一个公开维护元数据(Iof-maint)。Iof-maint是一个与工业知识库(IOF Core)对齐的模块化元数据,包含20个类和2个关系。它提供了一组用于各种实际数据驱动用例的维护特定术语。Iof-maint支持OWL DL推理,已在GitHub上进行了记录,并正在积极维护。在本文中,我们描述了Iof-maint参考元数据基于从多个应用 ontology 中提取共性概念的演变。
https://arxiv.org/abs/2404.05224
Chatbots can serve as a viable tool for preliminary depression diagnosis via interactive conversations with potential patients. Nevertheless, the blend of task-oriented and chit-chat in diagnosis-related dialogues necessitates professional expertise and empathy. Such unique requirements challenge traditional dialogue frameworks geared towards single optimization goals. To address this, we propose an innovative ontology definition and generation framework tailored explicitly for depression diagnosis dialogues, combining the reliability of task-oriented conversations with the appeal of empathy-related chit-chat. We further apply the framework to D$^4$, the only existing public dialogue dataset on depression diagnosis-oriented chats. Exhaustive experimental results indicate significant improvements in task completion and emotional support generation in depression diagnosis, fostering a more comprehensive approach to task-oriented chat dialogue system development and its applications in digital mental health.
聊天机器人可以成为通过与潜在患者进行具有交互性的对话进行初步抑郁诊断的可行工具。然而,在诊断相关的对话中混合了任务导向和闲聊,需要专业知识和同理心。这种独特的需求挑战了针对单一优化目标的傳統對話框架。为了应对这个问题,我们提出了一个专门针对抑郁诊断对话的創新本体定义和生成框架,结合了任务导向对话的可靠性和 empathy 相关的闲聊魅力。我们进一步将该框架应用于 D$^4,这是唯一一个关于抑郁诊断聊天数据的公共对话数据集。完整的实验结果表明,在抑郁诊断中,任务完成和情感支持生成的表现都有显著提高,促进了更全面的任务导向聊天对话系统开发和其在数字心理健康领域的应用。
https://arxiv.org/abs/2404.05012
Background. Endowing intelligent systems with semantic data commonly requires designing and instantiating ontologies with domain-specific knowledge. Especially in the early phases, those activities are typically performed manually by human experts possibly leveraging on their own experience. The resulting process is therefore time-consuming, error-prone, and often biased by the personal background of the ontology designer. Objective. To mitigate that issue, we propose a novel domain-independent approach to automatically instantiate ontologies with domain-specific knowledge, by leveraging on large language models (LLMs) as oracles. Method. Starting from (i) an initial schema composed by inter-related classes andproperties and (ii) a set of query templates, our method queries the LLM multi- ple times, and generates instances for both classes and properties from its replies. Thus, the ontology is automatically filled with domain-specific knowledge, compliant to the initial schema. As a result, the ontology is quickly and automatically enriched with manifold instances, which experts may consider to keep, adjust, discard, or complement according to their own needs and expertise. Contribution. We formalise our method in general way and instantiate it over various LLMs, as well as on a concrete case study. We report experiments rooted in the nutritional domain where an ontology of food meals and their ingredients is semi-automatically instantiated from scratch, starting from a categorisation of meals and their relationships. There, we analyse the quality of the generated ontologies and compare ontologies attained by exploiting different LLMs. Finally, we provide a SWOT analysis of the proposed method.
背景。为使智能系统获得语义数据,通常需要根据领域专业知识设计并实例化本领域的知识图谱。尤其是在最初阶段,这些活动通常由人类专家手动执行,可能还会利用他们自己的经验。因此, resulting process is therefore time-consuming, error-prone, and often biased by the personal background of the ontology designer. 目标。为了减轻这个问题,我们提出了一种新的、领域无关的方法来自动实例化具有领域特定知识的语义数据,通过利用大型语言模型(LLMs)作为预言者。方法。从(i)一个由相关类和属性组成的初始模式和(ii)一组查询模板开始,我们的方法多次查询LLM,并在其回复中生成类和属性的实例。因此,本语义图自动充满了领域特定知识,符合初始模式。因此,本语义图可以根据专家的需要和专业知识自动丰富多样实例,这些实例可以被视为保留、调整或丢弃。贡献。我们以一般方式形式阐述我们的方法,并在各种LLM上实例化它,同时还在一个具体案例研究中实例化它。我们在营养领域进行实验,从对餐食及其关系的分类开始,一个从零开始的餐食图谱 semi-自动实例化。在那里,我们分析了生成的语义图的质量,并比较了利用不同LLM获得的语义图的质量。最后,我们提供了所提议方法的SWOT分析。
https://arxiv.org/abs/2404.04108
The Artificial Intelligence Ontology (AIO) is a systematization of artificial intelligence (AI) concepts, methodologies, and their interrelations. Developed via manual curation, with the additional assistance of large language models (LLMs), AIO aims to address the rapidly evolving landscape of AI by providing a comprehensive framework that encompasses both technical and ethical aspects of AI technologies. The primary audience for AIO includes AI researchers, developers, and educators seeking standardized terminology and concepts within the AI domain. The ontology is structured around six top-level branches: Networks, Layers, Functions, LLMs, Preprocessing, and Bias, each designed to support the modular composition of AI methods and facilitate a deeper understanding of deep learning architectures and ethical considerations in AI. AIO's development utilized the Ontology Development Kit (ODK) for its creation and maintenance, with its content being dynamically updated through AI-driven curation support. This approach not only ensures the ontology's relevance amidst the fast-paced advancements in AI but also significantly enhances its utility for researchers, developers, and educators by simplifying the integration of new AI concepts and methodologies. The ontology's utility is demonstrated through the annotation of AI methods data in a catalog of AI research publications and the integration into the BioPortal ontology resource, highlighting its potential for cross-disciplinary research. The AIO ontology is open source and is available on GitHub (this https URL) and BioPortal (this https URL).
人工智能知识图谱(AIO)是一个对人工智能(AI)概念、方法和它们之间相互关系的系统化。AIO是通过手动策展开发起来的,并得到了大型语言模型(LLMs)的额外帮助。它旨在通过提供一个全面涵盖AI技术的技术和道德方面的框架,来应对AI领域快速变化的地形。AIO的主要受众包括AI研究人员、开发人员和教育者,他们寻求在AI领域使用标准化的术语和概念。AIO围绕六个顶级分支展开:网络、层、功能、LLMs、预处理和偏见,每个分支都旨在支持AI方法的模块化组合,并促进对深度学习架构和AI伦理问题的更深入理解。AIO的开发利用了知识图谱开发工具(ODK)进行创建和维护,其内容通过AI驱动的策展支持进行动态更新。这种方法不仅确保了AIO在AI快速发展的大背景下保持其相关性,而且显著提高了研究人员、开发人员和教育者的使用价值,通过简化新AI概念和方法的集成来提高其实用性。AIO的知识有用性通过将AI方法数据注释在AI研究出版物目录中,并将其集成到BioPortal元数据资源中,突出其跨学科研究潜力得到了充分证明。AIO是开源的,可以在GitHub(https://github.com)和BioPortal(https://biportal.org)上获取。
https://arxiv.org/abs/2404.03044
Social media is an easy-to-access platform providing timely updates about societal trends and events. Discussions regarding epidemic-related events such as infections, symptoms, and social interactions can be crucial for informing policymaking during epidemic outbreaks. In our work, we pioneer exploiting Event Detection (ED) for better preparedness and early warnings of any upcoming epidemic by developing a framework to extract and analyze epidemic-related events from social media posts. To this end, we curate an epidemic event ontology comprising seven disease-agnostic event types and construct a Twitter dataset SPEED with human-annotated events focused on the COVID-19 pandemic. Experimentation reveals how ED models trained on COVID-based SPEED can effectively detect epidemic events for three unseen epidemics of Monkeypox, Zika, and Dengue; while models trained on existing ED datasets fail miserably. Furthermore, we show that reporting sharp increases in the extracted events by our framework can provide warnings 4-9 weeks earlier than the WHO epidemic declaration for Monkeypox. This utility of our framework lays the foundations for better preparedness against emerging epidemics.
社交媒体是一个易于访问的平台,提供关于社会趋势和事件的及时更新。关于传染病相关事件(如感染、症状和社会互动)的讨论对于在疫情爆发期间进行政策制定至关重要。在我们的工作中,我们首创利用事件检测(ED)方法来更好地准备和早期预警即将到来的任何传染病。为此,我们构建了一个由七种疾病无关的事件类型组成的流行事件本体论,并构建了一个人类标注的Twitter数据集SPEED,以关注COVID-19大流行。实验表明,基于COVID的ED模型训练可以有效地检测出三种种类未见过的猴痘、登革热和黄热病等三个新的疫情;而基于现有ED数据集训练的模型则完全无法达到这种效果。此外,我们还证明了通过我们的框架提取的事件报告可以提前4-9周向世界卫生组织(WHO)颁布的疫情声明提供警告。这种基于我们的框架的更好的预防新兴传染病准备工作的基础。
https://arxiv.org/abs/2404.01679
Recognizing failures during task execution and implementing recovery procedures is challenging in robotics. Traditional approaches rely on the availability of extensive data or a tight set of constraints, while more recent approaches leverage large language models (LLMs) to verify task steps and replan accordingly. However, these methods often operate offline, necessitating scene resets and incurring in high costs. This paper introduces Recover, a neuro-symbolic framework for online failure identification and recovery. By integrating ontologies, logical rules, and LLM-based planners, Recover exploits symbolic information to enhance the ability of LLMs to generate recovery plans and also to decrease the associated costs. In order to demonstrate the capabilities of our method in a simulated kitchen environment, we introduce OntoThor, an ontology describing the AI2Thor simulator setting. Empirical evaluation shows that OntoThor's logical rules accurately detect all failures in the analyzed tasks, and that Recover considerably outperforms, for both failure detection and recovery, a baseline method reliant solely on LLMs.
在机器人领域,在任务执行过程中识别失败并实施恢复程序是非常具有挑战性的。传统方法依赖于大量数据或一组约束条件的可用性,而更现代的方法则利用大型语言模型(LLMs)来验证任务步骤并相应地重新规划。然而,这些方法通常需要离线操作,导致场景重置并产生高昂的成本。本文介绍了一个名为Recover的神经符号框架,用于在线故障识别和恢复。通过整合语义信息、逻辑规则和基于LLM的计划器,Recover利用符号信息增强LLMs生成恢复计划的能力,并降低相关成本。为了在模拟厨房环境中展示我们方法的性能,我们引入了OntoThor,一个描述AI2Thor仿真器设置的语义论。实证评估表明,OntoThor的逻辑规则准确地检测了分析任务中的所有故障,而Recover在故障检测和恢复方面都显著优于仅依赖LLM的基线方法。
https://arxiv.org/abs/2404.00756
Cooperative manufacturing is a new trend in industry, which depends on the existence of a collaborative robot. A collaborative robot is usually a light-weight robot which is capable of operating safely with a human co-worker in a shared work environment. During this cooperation, a vast amount of information is exchanged between the collaborative robot and the worker. This information constructs the cooperative manufacturing knowledge, which describes the production components and environment. In this research, we propose a holonic control solution, which uses the ontology concept to represent the cooperative manufacturing knowledge. The holonic control solution is implemented as an autonomous multi-agent system that exchanges the manufacturing knowledge based on an ontology model. Ultimately, the research illustrates and implements the proposed solution over a cooperative assembly scenario, which involves two workers and one collaborative robot, whom cooperate together to assemble a customized product.
合作制造是一种新兴的行业趋势,它依赖于合作机器人的存在。合作机器人通常是一种轻量级的机器人,能在共享工作环境中安全地与人类同事操作。在合作过程中,大量的信息在合作机器人和工人之间进行交换。这些信息构成了合作制造知识,描述了生产组件和环境。在这项研究中,我们提出了一个holonic控制解决方案,该方案使用本体论概念来表示合作制造知识。holonic控制解决方案被实现为基于本体模型的自主多代理系统,该系统基于本体模型交换制造知识。最后,研究在合作组装场景中展示了并实现了所提出的解决方案,该场景涉及两个工人和一个合作机器人,他们一起合作组装定制产品。
https://arxiv.org/abs/2404.00341
Modeling-Simulation-Optimization workflows play a fundamental role in applied mathematics. The Mathematical Research Data Initiative, MaRDI, responded to this by developing a FAIR and machine-interpretable template for a comprehensive documentation of such workflows. MaRDMO, a Plugin for the Research Data Management Organiser, enables scientists from diverse fields to document and publish their workflows on the MaRDI Portal seamlessly using the MaRDI template. Central to these workflows are mathematical models. MaRDI addresses them with the MathModDB ontology, offering a structured formal model description. Here, we showcase the interaction between MaRDMO and the MathModDB Knowledge Graph through an algebraic modeling workflow from the Digital Humanities. This demonstration underscores the versatility of both services beyond their original numerical domain.
建模-仿真-优化工作流程在应用数学中发挥着重要作用。为了回应这一需求,数学研究数据倡议(MaRDI)开发了一个全面记录这种工作流程的FAIR和机器可解释模板。MaRDMO,一个用于研究数据管理组织器的插件,使来自各种领域的科学家在MaRDI门户上使用MaRDI模板轻松地记录和发布他们的工作流程。这些工作流程的核心是数学模型。MaRDI用MathModDB本体解决这个问题,提供了一个结构化的形式化模型描述。在这里,我们通过数字人文领域的算术建模工作流程展示了MaRDMO和MathModDB知识图之间的互动。这个演示进一步证明了这两个服务的多样性和超出它们原有数值领域的应用范围。
https://arxiv.org/abs/2403.17778
We consider the problem of finding plausible knowledge that is missing from a given ontology, as a generalisation of the well-studied taxonomy expansion task. One line of work treats this task as a Natural Language Inference (NLI) problem, thus relying on the knowledge captured by language models to identify the missing knowledge. Another line of work uses concept embeddings to identify what different concepts have in common, taking inspiration from cognitive models for category based induction. These two approaches are intuitively complementary, but their effectiveness has not yet been compared. In this paper, we introduce a benchmark for evaluating ontology completion methods and thoroughly analyse the strengths and weaknesses of both approaches. We find that both approaches are indeed complementary, with hybrid strategies achieving the best overall results. We also find that the task is highly challenging for Large Language Models, even after fine-tuning.
我们将找到一个合理的知识,该知识从给定的本体中缺失,作为研究税目扩展任务的一般化。一种研究方法将这个问题视为自然语言推理(NLI)问题,因此它依赖于语言模型捕获到的知识来确定缺失的知识。另一种方法使用概念嵌入来确定不同概念之间的共同点,并从基于类别的归纳模型中获得灵感。这两种方法在直觉上是互补的,但它们的有效性尚未进行比较。在本文中,我们引入了一个评估本体完成度方法的基准,并深入分析两种方法的优缺点。我们发现,两种方法确实都是互补的,而混合策略实现的最佳整体结果。我们还发现,对于大型语言模型来说,即使经过微调,这个任务仍然具有很高难度。
https://arxiv.org/abs/2403.17216
Concept embeddings offer a practical and efficient mechanism for injecting commonsense knowledge into downstream tasks. Their core purpose is often not to predict the commonsense properties of concepts themselves, but rather to identify commonalities, i.e.\ sets of concepts which share some property of interest. Such commonalities are the basis for inductive generalisation, hence high-quality concept embeddings can make learning easier and more robust. Unfortunately, standard embeddings primarily reflect basic taxonomic categories, making them unsuitable for finding commonalities that refer to more specific aspects (e.g.\ the colour of objects or the materials they are made of). In this paper, we address this limitation by explicitly modelling the different facets of interest when learning concept embeddings. We show that this leads to embeddings which capture a more diverse range of commonsense properties, and consistently improves results in downstream tasks such as ultra-fine entity typing and ontology completion.
概念嵌入提供了一个实用且高效的机制,将常识知识注入到下游任务中。其核心目的通常不是预测概念本身的同义词,而是识别共同点,即具有共同兴趣的一组概念。这些共同点是归纳推理的基础,因此高质量的概念嵌入可以使学习更容易和更稳健。然而,标准嵌入主要反映了基本的分类范畴,因此它们不适合寻找指涉更具体方面(例如物体颜色或它们所使用的材料)的共同点。在本文中,我们通过明确建模学习概念嵌入时的不同兴趣方面,从而解决了这一局限。我们证明了这种方法导致具有更丰富多样性的概念嵌入,并且在下游任务(如超细实体类型和本体完成)中始终如一地提高结果。
https://arxiv.org/abs/2403.16984
Much of human knowledge in cybersecurity is encapsulated within the ever-growing volume of scientific papers. As this textual data continues to expand, the importance of document organization methods becomes increasingly crucial for extracting actionable insights hidden within large text datasets. Knowledge Graphs (KGs) serve as a means to store factual information in a structured manner, providing explicit, interpretable knowledge that includes domain-specific information from the cybersecurity scientific literature. One of the challenges in constructing a KG from scientific literature is the extraction of ontology from unstructured text. In this paper, we address this topic and introduce a method for building a multi-modal KG by extracting structured ontology from scientific papers. We demonstrate this concept in the cybersecurity domain. One modality of the KG represents observable information from the papers, such as the categories in which they were published or the authors. The second modality uncovers latent (hidden) patterns of text extracted through hierarchical and semantic non-negative matrix factorization (NMF), such as named entities, topics or clusters, and keywords. We illustrate this concept by consolidating more than two million scientific papers uploaded to arXiv into the cyber-domain, using hierarchical and semantic NMF, and by building a cyber-domain-specific KG.
大量的人类知识都封装在日益增长的科学论文中。随着文本数据的不断扩展,文档组织方法变得越来越重要,以便从大型文本数据集中提取潜在的具有行动意义的信息。知识图谱(KGs)作为一种存储事实信息的结构化方式,提供了明确的、可解释的知识,包括网络安全科学文献中的领域特定信息。构建KG从科学文献中的一大挑战是提取语义信息。在本文中,我们解决了这个问题,并引入了一种从科学论文中提取结构化语义的方法。我们在网络安全领域演示了这一概念。KG的一个模式代表来自论文的可观察信息,如它们所发表的分类或作者。另一个模式揭示了通过分层和语义非负矩阵分解(NMF)提取的潜在(隐藏)文本模式,例如命名实体、主题或聚类,以及关键词。我们通过使用分层和语义NMF将arXiv上超过2000万篇科学论文汇总到网络安全领域,并构建了一个网络安全领域特定的KG,来阐明这一概念。
https://arxiv.org/abs/2403.16222
This paper explores the integration of Large Language Models (LLMs) such as GPT-3.5 and GPT-4 into the ontology refinement process, specifically focusing on the OntoClean methodology. OntoClean, critical for assessing the metaphysical quality of ontologies, involves a two-step process of assigning meta-properties to classes and verifying a set of constraints. Manually conducting the first step proves difficult in practice, due to the need for philosophical expertise and lack of consensus among ontologists. By employing LLMs with two prompting strategies, the study demonstrates that high accuracy in the labelling process can be achieved. The findings suggest the potential for LLMs to enhance ontology refinement, proposing the development of plugin software for ontology tools to facilitate this integration.
本文探讨了将诸如GPT-3.5和GPT-4等大型语言模型(LLMs)融入元语优化过程的具体内容,特别是关注OntoClean方法论。OntoClean方法对于评估元语的质量至关重要,它包括将元属性分配给类和验证一组约束的两个步骤。在实践中,手动进行第一步证明很难,因为需要哲学专业知识,不同元理论家之间存在分歧。通过采用具有两种提示策略的LLM,这项研究证明了在实验室中实现高准确标注过程是可能的。研究结果表明,LLM可以在元语优化中提高准确性,建议为元语工具开发插件软件,以促进这一整合。
https://arxiv.org/abs/2403.15864
Social Media and Internet have the potential to be exploited as a source of opinion to enrich Decision Making solutions. Crowd Decision Making (CDM) is a methodology able to infer opinions and decisions from plain texts, such as reviews published in social media platforms, by means of Sentiment Analysis. Currently, the emergence and potential of Large Language Models (LLMs) lead us to explore new scenarios of automatically understand written texts, also known as natural language processing. This paper analyzes the use of ChatGPT based on prompt design strategies to assist in CDM processes to extract opinions and make decisions. We integrate ChatGPT in CDM processes as a flexible tool that infer the opinions expressed in texts, providing numerical or linguistic evaluations where the decision making models are based on the prompt design strategies. We include a multi-criteria decision making scenario with a category ontology for criteria. We also consider ChatGPT as an end-to-end CDM model able to provide a general opinion and score on the alternatives. We conduct empirical experiments on real data extracted from TripAdvisor, the TripR-2020Large dataset. The analysis of results show a promising branch for developing quality decision making models using ChatGPT. Finally, we discuss the challenges of consistency, sensitivity and explainability associated to the use of LLMs in CDM processes, raising open questions for future studies.
社交媒体和互联网具有成为意见来源的潜力,以丰富决策解决方案。众包决策(CDM)是一种能够通过情感分析从社交媒体平台上的普通文本中推断观点和决策的 methodology。目前,大型语言模型的出现和潜在可能性引发了我们对自动理解书面文本(自然语言处理)的新场景的探索。本文分析基于提示设计策略使用 ChatGPT 协助 CDM 过程以提取观点和做出决策。我们将 ChatGPT 集成到 CDM 过程中,作为一种灵活的工具,推断文本中表达的观点,为决策模型提供数值或语言评价。我们还包括一个多标准决策场景和一个类别本体,用于定义标准。我们还考虑 ChatGPT 是一个端到端 CDM 模型,能够提供对替代方案的一般意见和评分。我们在 TripAdvisor 和 TripR-2020Large 数据集上进行实证研究。分析结果表明,使用 ChatGPT 可以帮助开发质量决策模型。最后,我们讨论了使用 LLMs 在 CDM 过程中存在的挑战、敏感性和可解释性,提出了未来研究的开放问题。
https://arxiv.org/abs/2403.15587