re-include mdq context file after merge
include mdq composite score
remove root node in xpath (for some reason this wasn't working on mn-demo-8 but was locally).
Merge the code for rdfxml subprocessor from d1_cn_index_processor to metacat.
include v1 in the run mdq formatId
Add the prefix "prov_" the solr fields.
adjust color-based score calculations to match UI and add up to total result count
add indexing for scores based on successful checks by check.type
use dateConverter for run timestamp indexing
include initial MDQ run processing in metacat-index
Use a new class to overwrite the class RdfXmlSubprocessor in d1-processor since that one has a method to use solr http server directly.
merge changes from d1 indexing lib
Process the noaa variant of isotc211.
Add the file from d1_cn_index_processor.
Add a new copy from d1_cn_index_processor.
Add beans for the iso index.
Add a bean file for the iso index.
subclass AnnotatorSubprocessor for use in metacat-index (uses embedded solr server and solrj for retrieving/merging existing documents).
add fileName, mediaType and mediaTypeProperties to solr schema and v2 system metadata processor
refactor v2 context bean to use the v1 pattern used in metacat
include seriesId in solr schema and context file (v2 system metadata)
add missing quotation mark
fix xpath from CN changes for isPublic. https://redmine.dataone.org/issues/7374
include hierarchical permissions when evaluating isPublic during indexing. https://redmine.dataone.org/issues/7374
Index science metadata fields for the Dublin Core Extended metadata format. - Use d1_cn_index_processor 1.4.5 in metacat-index and update beans with new dcx subprocessor and xsi namespace
merge CN annotation context files to metacat (MN) to support semantic index fields.
use new IDocumentDeleteSubprocessors to handle clean-up of annotation index fields when annotations are removed.
move RDF/XML subprocessor and example configuration with SPARQL query to the cn index project from metacat so that it can be used by prov team when indexing ProvONE models in ORE documents
add fieldsToMerge property for annotation updates
update classes and context files that use cn-index-processor classes. allowing document subprocessors to be less tied to XML.
use sparql field and triple store from cn-index_processor (refactor). include annotatorSubprocessor for testing in metacat-index
let metacat-index lookup annotations for indexing rather than the metacat "reindex" action.
When indexing annotations from RDFs, use the doc id to access the system metadata, not the model name since they are not always the same.
Add PROV relationships to the Solr schema. Populate the fields using the RdfXmlSubprocessor
Add wasDerivedFrom field to the Solr schema and use Sparql query to retrieve the value from the RDF
update to use v2 types for indexing
include ORE formatId as handled by the RDF subprocessor and index prov:wasDerivedFrom field where it exists in the RDF model. https://projects.ecoinformatics.org/ecoinfo/issues/6548
Localized the file which doesn't have the bean for dataUrl.
Remove the reference to the bean eml.fileID.
Remove the bean named eml.fileID which used the ResolveSolrField class.
calculate geohash_3 to three places (typo)
use NSEW for the bounding box geohash calculation from EML - all versions
Add beans to support geohashes
augment annotation indexing test/sample to include orcid annotation. https://projects.ecoinformatics.org/ecoinfo/issues/6267https://projects.ecoinformatics.org/ecoinfo/issues/6423
include characteristic_sm field with SPARQL query
switch to index standard since it is more likely we will be able to determine this from our existing EML attribute information. https://projects.ecoinformatics.org/ecoinfo/issues/6253
switch to the OpenAnnotation (OA) model for annotating datapackages with measurements/characteristics (semtools)
use rangeOfDates | singleDateTime to populate the beginDate and endDate index fields. https://projects.ecoinformatics.org/ecoinfo/issues/6285
index singleDateTime value into both begin and end date fields in solr. https://projects.ecoinformatics.org/ecoinfo/issues/6285
allow indexing of RDF documents - provide a sparql query that will return values for the field name. Using measurement_sm initially (a dynamic multivalued solr field). https://projects.ecoinformatics.org/ecoinfo/issues/6253
Remove those files. They will be get from the d1_cn_index_processor 1.2.0 jar.
Add a patch for d1_cn_index_processor 1.1.2 version. So it can index taxon information. Those files will overwrite the ones in the d1_cn_index_process-1.1.2.jar.
Change the parameters order of the constructor. We maybe reuse some code from d1_cn_processor.
Remove the solr-home to the metacat-common.
Remove the file and it will got from the D1.
Remove the application context files (except the resource map one). It will use the ones in the d1_cn_index_processor.
Change the date format.Remove the replication part of log4j.
Add a log4j properties file.
Add a file to specify the log4j as the logger.
Use the new class to handle the resource map.
Add the necessary files for the solr home.
Move this file to the solr home and rename it to the schema.xml.
Add the solr home directory.
Remove this obsoleted file.
Add junit test class to test the SolrIndex class.
Add insert method for SolrIndex.
Figured out to read beans from the configuration file.