comparison scripts/migration/migration_instructions.txt @ 10:a50cf11e5178

Rewrite LGDataverse completely upgrading to dataverse4.0
author Zoe Hong <zhong@mpiwg-berlin.mpg.de>
date Tue, 08 Sep 2015 17:00:21 +0200
parents
children
comparison
equal deleted inserted replaced
9:5926d6419569 10:a50cf11e5178
1 Migration steps:
2
3 Assumptions:
4
5 - DVN 3.6 networkAdmin has id = 1
6 - Dataverse 4.0 admin has id = 1 (created by setup-all.sh script)
7
8
9 Pre steps (contained in the migration_presteps document):
10
11 -7. Make a copy of the production db, and point an app server to it
12 -6. (if there is any data that will fail validation, run scrubbing script - this will need to be custom per installation)
13 -5.9 run duplicate user scrubbing scripts
14 -5.8 run users as emails scripts
15 -5. Export DDI files from 3.6 copy for all datasets to be migrated
16 (this now includes exporting non-released versions - presteps doc. updated)
17 -4. Create copies of tables in 3.6 database for migrated data
18 -3. Run pg dump to extract tables copies
19 -2. Import copied tables into 4.0 database
20 -1. Run offsets on _dvn3_tables in the 4.0 DB
21
22 Migration:
23
24 1. run migrate_users.sql script
25 2. run migrate_dataverses.sql script
26 2a. migrate preloaded customizations
27 3. run custom_field_map.sql script (this must be updated to contain the custom field mappings specific to
28 the migration source installation.)
29 4. run dataset APIs: execute the following HTTP request on the Dataverse 4.0 application to initiate dataset migration:
30
31 http://<hostname>/api/batch/migrate?path=<parent directory of DDI files>&key=<Dataverse Admin API Key>
32
33 This will return a success message and begin an asynchronous migration job - the status of the job is viewable in the import-log file
34 in the Glassfish logs directory.
35
36 5. run migrate_datasets.sql script (post migration scrubbing)
37 6. run files script:
38
39 a. On the *destination* (4.0) server, step 1
40 run the script, and save the output:
41
42 ./files_destination_step1_ > migrated_datasets.txt
43
44 b. On the *source* (3.6) server -
45 run the script on the input produced in a.,
46 save the sql output:
47
48 ./files_source_ < migrated_datasets.txt > files_import.sql
49
50 (the script will also produce the output file packlist.txt,
51 to be used in step d., copying of physical files)
52
53 c. On the destination server, import the sql produced in b.:
54
55 psql <... params ...> -f files_import.sql
56
57 d. Package the files on the source server:
58
59 [TODO: there will be a script for this too]
60
61 as of now - can be done manually, by tarring the files listed
62 in packlist.txt
63
64 e. Unpack the files packaged above in the files directory on
65 the destination server.
66
67 7. run migrate_permissions.sql script (may need to delete some duplicates)
68 8. run migrate_links.sql script
69
70 10. reset sequences
71
72 11. (when ready for users to log in) add user passwords
73
74 __________________________________________________
75
76 Still to be migrated:
77 - Guestbook / stats
78
79
80 __________________________________________________
81
82 Not being migrated (verify?):
83 -- Study Comments
84 -- File Access requests
85 -- Classifications
86 -- Study locks
87 -- VDCNetworkStats (generated data)
88
89