add simple test for the IndexEventDAO class -- adding, removing, listing events in the DB table. https://projects.ecoinformatics.org/ecoinfo/issues/5944
Add the code that only the ids with the correct system metadata modification time will be added to the index queue.
Use the hazelcast event log.
Add code to get and set the last process date.
merging upgrade scripts from 2.0 branch to trunk. https://redmine.dataone.org/issues/3847
scripts for 2.1.0 upgrade
upgrade to Metacat 2.1.0 on the trunk. This includes a new index_event table for storing indexing events that need to be reprocessed. https://projects.ecoinformatics.org/ecoinfo/issues/5944
stub for storing IndexEvent objects in Metacat (from metacat-index processing). https://projects.ecoinformatics.org/ecoinfo/issues/5944
move IndexEvent into metacat-common. Perparation for Metacat responding to events and writing them to a persistent store. https://projects.ecoinformatics.org/ecoinfo/issues/5944
do not force a get() during refresh (causing EML-defined data access rules to be lost when inserting EML docs about data files). note that this reverses a change that was meant to trigger indexing, but now we are using a new queue to share index events with metacat-index and so should not be necessary.
do not use tmp file to return an inputstream on read() operations - just read from the file we already have. https://projects.ecoinformatics.org/ecoinfo/issues/6009
use standard File.createTempFile() method for uploaded data files and delete them when we are done with them. https://projects.ecoinformatics.org/ecoinfo/issues/6008
correctly delete data file when we are done with it. https://projects.ecoinformatics.org/ecoinfo/issues/6007
include filename in the filepart part. https://projects.ecoinformatics.org/ecoinfo/issues/6007
send the original filename in the upload() method since that is supported by the Metacat API. https://projects.ecoinformatics.org/ecoinfo/issues/6007
remove the unique string when generating data file metadata. https://projects.ecoinformatics.org/ecoinfo/issues/6007
use File::Temp to write data files in registry. https://projects.ecoinformatics.org/ecoinfo/issues/6007
correct regex for whitespace in D1 identifier.
refactor IndexEventLog a bit to simplify type/action information. prep for serializing IndexEvent objects to Metacat. https://projects.ecoinformatics.org/ecoinfo/issues/5944
remove serial number from indexeventlog - it is not used elsewhere in the api. https://projects.ecoinformatics.org/ecoinfo/issues/5944
correct spelling for index.eventlog.classname property
use an independent ISet<SystemMetadata> structure to communicate objects that should be indexed by metacat-index. https://projects.ecoinformatics.org/ecoinfo/issues/5943
consolidate SystemMetadata map retrieval in preparation for using a different structure for objects to index.
adding ability to remove event from the [error] queue.
do not create solr-home if there is no template to compy into that directory (need to be able to create it later if/when someone decides to use and deploy metacat-index). https://projects.ecoinformatics.org/ecoinfo/issues/6006
do not attempt to copy solr-home template from metacat-index webapp if it does not exist. This would be in cases where metacat-index is not deployed. https://projects.ecoinformatics.org/ecoinfo/issues/6006
Add code to implment set and get the last processed date.
It will make the index only for those objects which were modified after the marked time.
Add set and get the lastprocessedDate in the IndexEventLog.Remove the code to write the successful event.
Add the dataone repository.
The "war" target will build the metacat-index.war as well.
Log the timed index jobs.
Add the code to log the failed events.
Add a temporary file log for debugging.
Use commons-io 2.4
Add a new property for the index event log class name.
Add a serial number for the event.Add method to set events to be archived.
Add a new class variable - isArchived for class IndexEvent.
Update the documentation about those classes.
Add a event and eventlog for the index.
Use the identifier set to get the list of ids in the member node.
The returned ISet should be Identifier.
Add code to test getIdentifierSet method.
Add method to get identifier set.
Add a new property to specify the interval of a Timer to run the thread generating solr index.
Set up a Timer to run the regenerating solr index task periodically.
Use the ";" as the seperator to replace "," in the resource name spaces.
Add code to handle delete data package information when delete a pid in the solr index.
Add two static methods to get the SystemMetadata and data object InputStream for the specified id.
Change the code since the ApplicationController's constructor was changed.
Add code to check if the metacat.properties is available.
If solr is not enabled, it would not be running.
Solr will be enabled if it is in the db.enabledEngines.
Use ";" to separate db.enabledEngines.
Use the ";" as the seperator for properties.
do not require PortalCertificateManager be configured. Fix NPE because session was not created when using old sessionid-based authentication. https://projects.ecoinformatics.org/ecoinfo/issues/5942
change the waiting time to 10 seconds and attempts to 600 for the hazelcast.
Use another thread in the Servlet init method to wait hazelcast.
Make the target init depend on build-metacat-common.
Put the waiting mechanism for the hazelcast at the first place.
handle client certificates, portal certificates and jsessionid as three ways to prove you are an uthenticated user. https://projects.ecoinformatics.org/ecoinfo/issues/5942
Use some contants from the EnabledQueryEngines.
Temporarily remove the code to disable solr engine if it isn't listed in the property file.
exlude /lib/maven from the war file
If the solr engine is disabled, the metacat index will do nothing.
Use the new name of a method.
Rename a method to isEnabled.
Updated documentation, and added modification date to the sitemap index file entries.
Remove unused import.
Mofdified Sitemap class to also generate the sitemap index file that is needed when more than one sitemap file is provided.
Remove the junit test for an obsoleted class.
Remove the obsoleted class.
Add a junit test class for EnabledQueryEngines.
Add a test base class.
Add a class to repsent the enabled engine list.
Change the junit version to 4.8
use ContentTypeInputStream interface (and ByteArray implementation) to specify the desired content-type of the InputStream returned by MN.query().
Add the new property dbquery.enabledEngines=pathquery,solr.
load the evicted SM back into the map on a "Refresh" so that listeners hear the update. (metacat-index, for example)
set the dataone.hazelcast.location.clientconfig property to be the local config file. It will be used in the d1_cn_index_processor
use DataONE cilogon skin so that we are not prompted for certificate duration (and to make it more cohesive as a prototype using the Metacat DataONE skin
switch back to log4j statements now that I am sure certificate delegation is working.
add link to login or switch user
add simple page to show logged-in user subject DN
use System.out.println until the oa4mp logging issue is resolved.
Read the waiting time and maximum attempts from the metacat.properties file.
Add the waiting time and max attempts for regenerating the index.
use filesystem store. in-memory store seems to get into deadlock with multiple environments accessing it.
add logging for portal certificate look up process.
Add some new properties for the test.
use relative path for oa4mp_client.xml (within servlet context). https://projects.ecoinformatics.org/ecoinfo/issues/5936
use in-memory storage for myproxy. at least for now.
Add a property containing the list of the resource map namespaces.
Index the none-resourcemap data objects first, then the resourcemap objects.
used pk8 format private key file name
use java 1.6 compliance for compilation
first pass at integrating CILogon/MyProxy certificates in Metacat. Configuration is specific to mn-demo-4.test.dataone.org for the time being (this will cause localhost deployments to fail webapp deployment). https://projects.ecoinformatics.org/ecoinfo/issues/5936
Updated Sitemap generation to use latest version of the sitemap protocol schemas.