# Utilities to read/write csv files import csv # Utilities to handle character encodings import unicodedata # Ordered Dicts from collections import OrderedDict import json import string # OPZIONAL IMPORTS # For timestamping/simple speed tests from datetime import datetime # Random number generator from random import * # System & command line utilities import sys # Json for the dictionary import json import_dir = '/Users/federicaspinelli/TEAMOVI/Parser/DATA/TOPONIMI/' export_dir = '/Users/federicaspinelli/TEAMOVI/Parser/DATA/TOPONIMI/' # Custom class to store URIs + related infos for the ontologies/repositories class RDFcoords: def __init__(self, uri, prefix, code = None): self.uri = uri self.prefix = prefix self.code = code # Repositories aspoCoords = RDFcoords('', 'dat:') placeCoords = RDFcoords('', 'pl:') oviCoords = RDFcoords('', 'ovi:') cidocCoords = RDFcoords('', 'crm:') tgnCoords = RDFcoords('', 'tgn:') repettiCoords = RDFcoords('', 'rpt:') schemaCoords = RDFcoords('', 'rdfs:') owlCoords = RDFcoords('', 'owl:') devCoords = RDFcoords('', 'dev:') # SKOS skosCoords = RDFcoords('', 'skos:') # prefLabelCoords = RDFcoords('', 'pref:') # narrowerCoords = RDFcoords('', 'narr:') # broaderCoords = RDFcoords('', 'bro:') # Basic functions for triples / shortened triples in TTL format def triple(subject, predicate, object1): line = subject + ' ' + predicate + ' ' + object1 return line def doublet(predicate, object1): line = ' ' + predicate + ' ' + object1 return line def singlet(object1): line = ' ' + object1 return line # Line endings in TTL format continueLine1 = ' ;\n' continueLine2 = ' ,\n' closeLine = ' .\n' def writeTTLHeader(output): output.write('@prefix ' + placeCoords.prefix + ' ' + placeCoords.uri + closeLine) output.write('@prefix ' + aspoCoords.prefix + ' ' + aspoCoords.uri + closeLine) output.write('@prefix ' + cidocCoords.prefix + ' ' + cidocCoords.uri + closeLine) output.write('@prefix ' + tgnCoords.prefix + ' ' + tgnCoords.uri + closeLine) output.write('@prefix ' + schemaCoords.prefix + ' ' + schemaCoords.uri + closeLine) output.write('@prefix ' + repettiCoords.prefix + ' ' + repettiCoords.uri + closeLine) output.write('@prefix ' + owlCoords.prefix + ' ' + owlCoords.uri + closeLine) output.write('@prefix ' + devCoords.prefix + ' ' + devCoords.uri + closeLine) output.write('@prefix ' + skosCoords.prefix + ' ' + skosCoords.uri + closeLine) output.write('\n') file = "toponimi_DEV" max_entries = 1000000000 with open(import_dir + file + '.csv', newline="") as csv_file, open( export_dir + file + '.ttl', 'w') as output: reader = csv.DictReader(csv_file) writeTTLHeader(output) first = True ii = 0 for row in reader: # The index ii is used to process a limited number of entries for testing purposes ii = ii + 1 tgnPlaceHolder = tgnCoords.prefix + row['TGN'] line = triple(tgnPlaceHolder, schemaCoords.prefix + 'type', 'Province') + closeLine output.write(line) line = triple(tgnPlaceHolder, skosCoords.prefix + 'narrower', 'Province') + closeLine output.write(line) line = triple(tgnPlaceHolder, skosCoords.prefix + 'broader', 'Province') + closeLine output.write(line) #placeHolders # devPlaceHolder = devCoords.prefix + row['ID RESTORE'] # line = triple(devPlaceHolder, # nsCoords.prefix + 'type', # cidocCoords.prefix + 'E53_Place') + closeLine # output.write(line) # label = string.capwords(row['TOPONIMO']) # line = triple(devPlaceHolder, schemaCoords.prefix + 'label', '\"' + label + '\"') + closeLine # output.write(line) # line = triple(devPlaceHolder, # cidocCoords.prefix + 'P1_is_identified_by', # '\"' + row['ID RESTORE'] + '\"') + closeLine # output.write(line) # if row['LAT'] != '' and row['LONG'] != '': # line = triple(devPlaceHolder, # cidocCoords.prefix + "P168_place_is_defined_by", # '\"' + row['LAT'] + ' , ' + row['LONG'] + '\"') + closeLine # output.write(line) # if row['TGN'] != '' and row['TGN'] != 'FALSE': # tgnPlaceHolder = tgnCoords.prefix + row['TGN'] # line = triple(devPlaceHolder, # owlCoords.prefix + 'sameAs', # tgnPlaceHolder) + closeLine # output.write(line) # line = triple(tgnPlaceHolder, # schemaCoords.prefix + 'label', # '\"' + label + ' (TGN)\"') + closeLine # output.write(line) # line = triple(tgnPlaceHolder, # cidocCoords.prefix + 'P1_is_identified_by', # '\"' + row['TGN'] + '\"') + closeLine # output.write(line) # if row['REPETTI'] != '' and row['REPETTI'] != 'FALSE': # repettiPlaceHolder = repettiCoords.prefix + row['REPETTI'] # line = triple(devPlaceHolder, # owlCoords.prefix + 'sameAs', # repettiPlaceHolder) + closeLine # output.write(line) # line = triple(repettiPlaceHolder, # schemaCoords.prefix + 'label', # '\"' + label + ' (REPETTI)\"') + closeLine # output.write(line) # line = triple(repettiPlaceHolder, # cidocCoords.prefix + 'P1_is_identified_by', # '\"' + row['REPETTI'] + '\"') + closeLine # output.write(line) output.write('\n') # # # Limit number of entries processed (if desired) if (ii > max_entries): break