only run ORE generation for EML docs -- no need to run this for all documents (yikes!)
use IdMan method to find docids that do not already have system metadata records -- this lets us re-run without re computing system metadata for every entry (in case the process is interrupted). I haven' been using this option because I wanted to continually regenerate all SM for everything in my test DBs, but we are so close to release that I want to get this in there.
for testing: limit and randomize the docs to generate metadata for
FOR TESTING ONLY: limit number of records to 100 so that we can get an estimate
update the memberNodeId in existing system metadata only after the register/update is successful with the CN -- we can avoid unneeded SM updates in cases when the register/update fails because we gave the CN bad info that it rejects. https://redmine.dataone.org/issues/2308
include member node id text field now that the CN is not assigning random Ids.https://redmine.dataone.org/issues/2308
1. lookup and use the guid when processing obsoletes/obsoletedBy entries -- had previously been assuming localId==guid but now that we have introduced DOIs as part of the Metacat upgrade process, we may have DOIs for the guid that map to localIds.2. base ORE guids on the localid of the data package they are describing and not on their DOI -- otherwise we might mash up the DOI prefix (or other id scheme that we are unaware of). By using resourceMapPreix + localId we are sure to have a valid localid and guid for the ORE map we create and add to the system
use updated authorization policies as discussed in:https://redmine.dataone.org/issues/2277andhttp://epad.dataone.org/20120131-authn-authz-questions
refactor D1-specific upgrade utilities into their own package
remove createAndInsertSystemMetadat() method that acts on a single localId -- incorporated this into the localId-list-based method.
refactor IdentityManager.createSystemMetadata(sm) to be insertSystemMetadata(sm) so that it is clear that this method inserts the SM object into the backing store. This differentiates it from the "generation" methods we use when we need to create SM about pre-existing objects or objects we get from non-D1 api calls.
generate SystemMetadata during D1 registration (not 2.0.0 upgrade). This process runs in a thread and updates a metacat.properties value when it is complete.
getMultipartParameters() outside of debug block -- thanks Mark Reyes @ CDL for catching this.
dataone configuration and registration enhancements:-include flag to disable D1 services, currently only the MN side enforces this-do not allow multiple registration attempts if we have just submitted and are awaiting Node verification by the CN.-do not allow configuration "bypass" if D1 settings have been configured previously....
use correct Collections import
Show "Update" button if this MemberNodeId is already registered with DataONE, otherwise use the "Register" label
match changes to MN service methods (return type as boolean)
updated d1 jars with latest libclient changes and objectformatcache use
Updated configuration documentation in admin guide for Metacat DataONE section. Changed links in configuration utility to point at the Admin guide.
Added new methods to generate a default replication policy based on properties from the metacat configuration. This is called during system metadata creation for objects that lack any system metadata.
Modify admin configuration to include default replication policy. Extensively revised the DataONE configuration page, including new wording for intro, improved tooltips throughout, new arrangement of sections, and other cosmetic changes.
Clean up warnings in class.
Remove ability to edit NodeID from D1 configuration page. Fix update of contactSubject and dataone.ore.generated property name.
include flag indicating that system metadata generation has completed (useful for independent long-running thread)
handle "BIN" objects so as to avoid repeated calls to lookup the non-existent ObjectFormat
do not wait for SM generation to complete during the upgrade -- this way the web UI wont hang for days. the process sets a metacat property when it is complete.
Fixed a bug the a hyper-link included the username/password input fields.
use RC-2 DataONE jars -- these are built from trunk still, but include the next tag naming convention
do not shutdown hazelcast -- it needs to be running after the upgrade process so that Metacat actually works.I think the newer version of HZ makes it so the threads are all released as needed.http://bugzilla.ecoinformatics.org/show_bug.cgi?id=5572
Commenting out the parts of the upgrade script that started to refer to EXID. At this point, the registration of EZID identifiers will be done out-of-band with respect to the upgrade.
use plain String parameter for {pid} instead of XML serialization of it.
remove {pid} from POST URL on CN.registerSystemMetadata()https://redmine.dataone.org/issues/2284
remove {pid} from POST URL on CN.create()https://redmine.dataone.org/issues/2284
remove {pid} from POST URL on MN.create()https://redmine.dataone.org/issues/2284
catch cases where the previous/next revision of objects have not had system metadata generated yet
create system metadata object if it wasn't found in HZ
adjust the width of the label suffix.
process systemMetadata from the docInfo string before writing to the database so that we guarantee guid-docid mapping exists before attempting to look it up.
Adjust the column width of the search result.
upgrade to hazelcast 1.9.4.6 so that threadpools are released when not needed (http://code.google.com/p/hazelcast/issues/detail?id=765).include ant target to run a specific main class (mostly for debugging)
use File.deleteOnExit() not a half hour timer thread to do it.
multithreaded implementation for processing docids for system metadata generation.need to investigate ant/junit running that deadlocks hazelcast (config?)
additional logging of the config file being used - seem to have thread locking on the xmlConfig use when running under ant/junit
calculate object size using the size on the file system rather than re-reading as an input stream.Now only EML document bytes will be read twice: once for the checksum and again for parsing out datapackage details
system metadata generation optionally skips entries that have already been generated (data size, checksum) but allows the latest EML that describes them to have the last word on object format
remove DML for parsing -- the D1 EML parser still uses DOM, so this may not be too big of a perfromance improvement
test harness for running system metadata generation outside of the upgrade process
include comment about KNB estimated time to run during upgrade:Total time: 20 minutes 58 seconds
only attempt to update date-like nodedata values.
use "test" to exercise upgrade code on staging DB.
include generate system metadata upgrade in the success flag
more clean up - reuse prepared statement for data update
look up nodedata values first, then update each one - trying to avoid out of memory exception.
eliminate the cross product that occurred when updating xml_access with a join
rollback processing Error change -- creates a loop on error. ugh
report processing errors after exceptions have been caught and recorded, otherwise the web UI is blank and there is no clue what happened unless you look in the logs.
semicolons!
fix a bug in MNodeService.replicate() where the checksum value was being compared to the computed checksum object, not its value.
use a temporary table to calculate the maximum revision for a given docid and use that when setting the accessfileid during upgrade. the query plan for the all-in-one statement must be brutal as it's been running for 4 hours at this point....
do not insert duplicate GUID entries when adding rows from the xml_revisions table
add "IF EXISTS" clause to identifier table drop in case it does not exist on the given deployment (as is the case on the KNB)
use UTC serialization for log entries so that the timestamp, not just the date, is preservedhttps://redmine.dataone.org/issues/2257
edited the configuration docs, swapped in new fully-configured screen shothttp://bugzilla.ecoinformatics.org/show_bug.cgi?id=5516
Update the MNodeServiceTest to test the validity of the node document returned by getCapabilities() by parsing it with the TypeMarshaller.
Update the D1Admin class to set the dataone.contactSubject property. I've added the property to the http request to be added to the JSP form, but for now am setting the property using the dataone.subject field value. Not sure if we want to expose the contact subject in the form yet or not.
marked the properties that are only editable manually in metacat.propertieshttp://bugzilla.ecoinformatics.org/show_bug.cgi?id=5516
Add the dataone.contactSubject property to be used when getCapabilities() returns the MN node instance.
In MN.getCapabilities(), the required contact subject was not being added to the node instance from the dataone properties. Add it in.
format xml/eml table -- added missing row delimiterhttp://bugzilla.ecoinformatics.org/show_bug.cgi?id=5516
some minimal editing on replication configuration -- bulk of it had been done in the Word doc before rst conversionhttp://bugzilla.ecoinformatics.org/show_bug.cgi?id=5516
generate ORE maps only once -- and persist the flag to the main backup properties so that subsequent Metacat upgrades remember this value.
use RC-1 Dataone jars
Added DOI generation to the 2.0.0 upgrade process. To succeed, this script must be run on a fresh 2.0.0 database, or on a 1.9.5 version database, as those are the only ways to get the needed foreign keys to be marked as deferrable. The identifier conversion must be turned on by setting correct properties in metacat.properties. See the comments in GenerateGlobalIdentifiers for details. By default, conversion is set to false in the properties file. If you want to convert an instance to use DOIs, be sure to set metacat.properties up BEFORE running through the Metacat configuration and database upgrade.
Refactoring classes that throw generic Exception class to throw their more specific subclasses so that new exceptions are not hidden behind generic messages. Makes debugging easier.
try to read the local document before making the localid->guid mapping (in cases where we fail to read the data locally like if it is referenced in an EML file but does not exist on this Metacat instance)
add note to enable ssl mod
-include instructions about enabling AJP connector in Tomcat config-use postgres 8.4-include hint to disable default Apache site (otherwise the knb config never gets used!)
include instructions for installing Sun's Java6 JDK since it is no longer in apt-get repos
Ensure we have the object and sysmeta params for MN.create(). We were getting a fatal SAX parsing error encapsulated in a ServiceFailure when a science metadata object param was null. Cut it off at the pass after parsing the MMP entity.
An example python script that uses the python client to loop through a list offiles, read them from disk, and insert them into metacat.
make it clear that the Apache config files are samples and may need to be modified for different servershttp://bugzilla.ecoinformatics.org/show_bug.cgi?id=5516
update system requirements to be more reasonable
use larger ("text") db field for guid in the xml_access.accessfileid column
use EML 2.1.1 tag as final tag for the schemahttp://bugzilla.ecoinformatics.org/show_bug.cgi?id=5532
use RELEASE_EML_UTILS_1_0_0 for EML style sheetshttp://bugzilla.ecoinformatics.org/show_bug.cgi?id=5532
comment our the Demis world map layer -- it was prompting for username/password to use the WMS
Use the Collections class from java.util.
Remove null field tests in the IdentifierManager class. Schema-level required fields are checked on serialization/deserialization using JibX during the REST resource handler classes. Other required fields are checked in MNodeService and CNodeService, higher in the stack.
For MNs that haven't set the archived flag to false on create(), set it here. Also, ensure that the CN sync code sets the authoritative and origin member node fields.
On MN.create(), set the archived flag to be false. This field isn't required in the schema, but is needed by the DataONE indexer once objects are sync'd.
use EML 2.1.1 RC4 tag before final tag (schema)http://bugzilla.ecoinformatics.org/show_bug.cgi?id=5532
use final tag for building with utilities (tags/UTILITIES_1_1_0)http://bugzilla.ecoinformatics.org/show_bug.cgi?id=5532
use final tag for building with ecogridhttp://bugzilla.ecoinformatics.org/show_bug.cgi?id=5532
-generate system meta for all docids, even those not originating on the server (replicas from the past)-generate ORE docs and download remote data only for those documents that originated on this server being upgraded.http://bugzilla.ecoinformatics.org/show_bug.cgi?id=5522
refactor generate system meta loop to the factory class -- to be reused in sysmeta and ORE generationhttp://bugzilla.ecoinformatics.org/show_bug.cgi?id=5522
When managing obsoletes/obsoletedBy system metadata fields, set the archived flag to false initially, and set it to true on system metadata for objects that a revision obsoletes.
do NOT generate ORE maps or download data when we do the initial System Metadata generation -- this is deferred until D1 registration.
make more generic so that a custom list of IDs can be passed in.