Mercurial > hg > LGDataverses
comparison scripts/migration/migration_instructions.txt @ 10:a50cf11e5178
Rewrite LGDataverse completely upgrading to dataverse4.0
| author | Zoe Hong <zhong@mpiwg-berlin.mpg.de> |
|---|---|
| date | Tue, 08 Sep 2015 17:00:21 +0200 |
| parents | |
| children |
comparison
equal
deleted
inserted
replaced
| 9:5926d6419569 | 10:a50cf11e5178 |
|---|---|
| 1 Migration steps: | |
| 2 | |
| 3 Assumptions: | |
| 4 | |
| 5 - DVN 3.6 networkAdmin has id = 1 | |
| 6 - Dataverse 4.0 admin has id = 1 (created by setup-all.sh script) | |
| 7 | |
| 8 | |
| 9 Pre steps (contained in the migration_presteps document): | |
| 10 | |
| 11 -7. Make a copy of the production db, and point an app server to it | |
| 12 -6. (if there is any data that will fail validation, run scrubbing script - this will need to be custom per installation) | |
| 13 -5.9 run duplicate user scrubbing scripts | |
| 14 -5.8 run users as emails scripts | |
| 15 -5. Export DDI files from 3.6 copy for all datasets to be migrated | |
| 16 (this now includes exporting non-released versions - presteps doc. updated) | |
| 17 -4. Create copies of tables in 3.6 database for migrated data | |
| 18 -3. Run pg dump to extract tables copies | |
| 19 -2. Import copied tables into 4.0 database | |
| 20 -1. Run offsets on _dvn3_tables in the 4.0 DB | |
| 21 | |
| 22 Migration: | |
| 23 | |
| 24 1. run migrate_users.sql script | |
| 25 2. run migrate_dataverses.sql script | |
| 26 2a. migrate preloaded customizations | |
| 27 3. run custom_field_map.sql script (this must be updated to contain the custom field mappings specific to | |
| 28 the migration source installation.) | |
| 29 4. run dataset APIs: execute the following HTTP request on the Dataverse 4.0 application to initiate dataset migration: | |
| 30 | |
| 31 http://<hostname>/api/batch/migrate?path=<parent directory of DDI files>&key=<Dataverse Admin API Key> | |
| 32 | |
| 33 This will return a success message and begin an asynchronous migration job - the status of the job is viewable in the import-log file | |
| 34 in the Glassfish logs directory. | |
| 35 | |
| 36 5. run migrate_datasets.sql script (post migration scrubbing) | |
| 37 6. run files script: | |
| 38 | |
| 39 a. On the *destination* (4.0) server, step 1 | |
| 40 run the script, and save the output: | |
| 41 | |
| 42 ./files_destination_step1_ > migrated_datasets.txt | |
| 43 | |
| 44 b. On the *source* (3.6) server - | |
| 45 run the script on the input produced in a., | |
| 46 save the sql output: | |
| 47 | |
| 48 ./files_source_ < migrated_datasets.txt > files_import.sql | |
| 49 | |
| 50 (the script will also produce the output file packlist.txt, | |
| 51 to be used in step d., copying of physical files) | |
| 52 | |
| 53 c. On the destination server, import the sql produced in b.: | |
| 54 | |
| 55 psql <... params ...> -f files_import.sql | |
| 56 | |
| 57 d. Package the files on the source server: | |
| 58 | |
| 59 [TODO: there will be a script for this too] | |
| 60 | |
| 61 as of now - can be done manually, by tarring the files listed | |
| 62 in packlist.txt | |
| 63 | |
| 64 e. Unpack the files packaged above in the files directory on | |
| 65 the destination server. | |
| 66 | |
| 67 7. run migrate_permissions.sql script (may need to delete some duplicates) | |
| 68 8. run migrate_links.sql script | |
| 69 | |
| 70 10. reset sequences | |
| 71 | |
| 72 11. (when ready for users to log in) add user passwords | |
| 73 | |
| 74 __________________________________________________ | |
| 75 | |
| 76 Still to be migrated: | |
| 77 - Guestbook / stats | |
| 78 | |
| 79 | |
| 80 __________________________________________________ | |
| 81 | |
| 82 Not being migrated (verify?): | |
| 83 -- Study Comments | |
| 84 -- File Access requests | |
| 85 -- Classifications | |
| 86 -- Study locks | |
| 87 -- VDCNetworkStats (generated data) | |
| 88 | |
| 89 |
