= Data Integration = == Integration of Cologne Data == === general === '''convertFaust''' converts faust-xml output into FMPXMLRESULT. ''' match_faust_arb''' contains the methods for matching the different object with the ARB data. === texts === First step is matching the texts. * convert text objects from faust to filemaker format. * import the texts into filemaker database faust-texte. * export relevant field for matching in this case only the number part of '''signatur''' is comparted with the numbers in '''Title''' in N_source_codex * match with '''match_faust_arb.doMatchTexts''' * import results into C_faust_arb_texte * manually check (layout -> C_faust_arb_texte) Relating the matched texts: * Layout N_source_codex_faust displayes links to the matched texts. Importing the new texts/ all non matching texts: * Import all faust-ids into N_source_codex, for each is a new empty data set created, original ID is saved in __import_from_faust * Script import_from_faust_text, imports selected fields from the faust-db into ARB * Script import_all_from_faust_text, loops through all selected datasets and does the import, if possible. Matching: * [ticket:25 Contents / Remarks of the matching docs have to be integrated?] * [ticket:26 needs an layout to do so] === recipes === * [ticket:21 Matching of recipies has to be done] * [ticket:18 Import non matching recipes ] (Import all recipies which can't be in ARB because the texts are not in ARB. (Seems to be only one ID:4229) * [ticket:22 Import remaining recipes] * first: import von faust-recepte: kommentar -> remark, key --> {{{__import_key__from_cologne}}}, Material --> material * second : import now : Folio_Seite --> location, c_faust_arb_texte:correct_arb --> ID_target, N_recipies:ID --> ID_source * Now export the now generated datasets (fields: ID_source, ID) * Now import the exported data into N_recipes, with ID_soruce matching ID und ID -> ID_recipes_sources * Now import transcription and translation: * Export transcription and Übersetzung mit ID_recep_ * Import in N_representation transkription --> transcription, Übersetzung --> translation, C:recipes_soruces:ID --> _tmp * Export the new datasets from there _tmp,ID * Import this into C_recipes_source, witch matching _tmp -> ID_source, ID -> ID_representation * Export from faust-lexikon.fp7 N_ingredient, N_recipes and key * Generate from the export file with splitExport mapping for C_recipes_ingredient. * Import this into C_recipes_ingredient * Import in N_appellation from faust-lexikon, key -> __import_from_cologne, name -> appelation * Export from N_appellation {{{ __import__from_cologne_ID }}} and ID * Import into C_recipes_ingredient matching {{{ __import__from_cologne_ID}}} ID --> ID_appellation * Import in N_lemma from faust-lexikon, key -> __import_from_cologne, lemma -> name * Export from N_Lemma {{{ __import__from_cologne_ID }}} and ID * Import into C_recipes_ingredient matching {{{ __import__from_cologne_ID}}} ID --> ID_lemma * Results need checking [ticket:28 Lemmas] and [ticket:29 Appellation] === lexikon === * [ticket:24 General: solves problem material vs. ingredient.] * First step: matching DB_glossary and ingredients. (inported all missing, and comments), [ticket:23 should be checked by Sylvie] * Import all Moderne Namen ";" separated into to entries in N_ingredient (datei: moderneNamen_id.csv), new field lexikon_id_cologne, this ID was created by a script (modernNameMatch) and is not in the FAUST db. [wiki:remarks Remarks]