- Introduced `test_nlp_extractor.py` with unit tests for the InstitutionExtractor, covering various extraction patterns (ISIL, Wikidata, VIAF, city names) and ensuring proper classification of institutions (museum, library, archive). - Added tests for extracted entities and result handling to validate the extraction process. - Created `test_partnership_rdf_integration.py` to validate the end-to-end process of extracting partnerships from a conversation and exporting them to RDF format. - Implemented tests for temporal properties in partnerships and ensured compliance with W3C Organization Ontology patterns. - Verified that extracted partnerships are correctly linked with PROV-O provenance metadata.
14 lines
388 B
Python
14 lines
388 B
Python
"""
|
|
Wikidata enrichment utilities for heritage institution data.
|
|
|
|
This module provides two-tier search strategies:
|
|
1. Primary: SPARQL queries (fast, structured)
|
|
2. Fallback: MediaWiki API search (slower, broader coverage)
|
|
"""
|
|
|
|
from .search import search_wikidata_mediawiki, search_wikidata_with_fallback
|
|
|
|
__all__ = [
|
|
'search_wikidata_mediawiki',
|
|
'search_wikidata_with_fallback'
|
|
]
|