Compare commits

...

2 Commits

Author SHA1 Message Date
b86d82fb87 Modified conllu_to_tei.py for multiple files. 2022-11-04 08:11:59 +00:00
e636be1dc2 Added NER + SRL to conllu_to_tei script 2022-11-03 09:31:41 +00:00
4 changed files with 155 additions and 35 deletions

2
.gitignore vendored
View File

@ -1,2 +1,4 @@
*.pyc *.pyc
venv venv
data
.idea

View File

@ -4,20 +4,34 @@ import sys
from lxml import etree from lxml import etree
from conversion_utils.jos_msds_and_properties import Converter, Msd
converter = Converter()
def translate_msd(msd_text, lang, lemma=None):
""" Translates msd using conversion_utils library. """
return converter.properties_to_msd(converter.msd_to_properties(Msd(msd_text, 'en'), 'sl', lemma),
'sl').code
class Sentence: class Sentence:
def __init__(self, _id, no_ud=False, system='jos'): def __init__(self, _id, no_ud=False, system='jos'):
self._id = _id self._id = _id
self.items = [] self.items = []
self.links = [] self.links = []
self.srl_links = []
self.no_ud = no_ud self.no_ud = no_ud
self.system = system self.system = system
def add_item(self, token, lemma, upos, upos_other, xpos, misc): def add_item(self, token, lemma, upos, upos_other, xpos, misc):
self.items.append([token, lemma, upos, upos_other, xpos, "SpaceAfter=No" in misc.split('|')]) no_space_after = 'SpaceAfter' in misc and misc['SpaceAfter'] == 'No'
ner = misc['NER'] if 'NER' in misc else 'O'
self.items.append([token, lemma, upos, upos_other, xpos, no_space_after, ner])
def add_link(self, link_ref, link_type): def add_link(self, link_ref, link_type):
self.links.append([link_ref, link_type]) self.links.append([link_ref, link_type])
def add_srl_link(self, link_ref, link_type):
self.srl_links.append([link_ref, link_type])
def as_xml(self, id_prefix=None): def as_xml(self, id_prefix=None):
if id_prefix: if id_prefix:
xml_id = id_prefix + '.' + self._id xml_id = id_prefix + '.' + self._id
@ -27,8 +41,24 @@ class Sentence:
set_xml_attr(base, 'id', xml_id) set_xml_attr(base, 'id', xml_id)
id_counter = 1 id_counter = 1
in_seg = False
sentence_base = base
for item in self.items: for item in self.items:
token, lemma, upos, upos_other, xpos, no_space_after = item token, lemma, upos, upos_other, xpos, no_space_after, ner = item
if ner[0] == 'B':
if in_seg:
sentence_base.append(base)
in_seg = True
base = etree.Element('seg')
base.set('type', 'name')
base.set('subtype', f'{ner[2:].lower()}')
elif ner[0] == 'O':
if in_seg:
sentence_base.append(base)
base = sentence_base
in_seg = False
if xpos in {'U', 'Z'}: # hmm, safe only as long as U is unused in English tagset and Z in Slovenian one if xpos in {'U', 'Z'}: # hmm, safe only as long as U is unused in English tagset and Z in Slovenian one
to_add = etree.Element('pc') to_add = etree.Element('pc')
@ -36,6 +66,7 @@ class Sentence:
to_add = etree.Element('w') to_add = etree.Element('w')
to_add.set('lemma', lemma) to_add.set('lemma', lemma)
xpos = translate_msd(xpos,'sl',lemma)
to_add.set('ana', 'mte:' + xpos) to_add.set('ana', 'mte:' + xpos)
if not self.no_ud: if not self.no_ud:
if upos_other != '_': if upos_other != '_':
@ -53,6 +84,11 @@ class Sentence:
base.append(to_add) base.append(to_add)
if in_seg:
sentence_base.append(base)
base = sentence_base
# depparsing linkGrp
link_grp = etree.Element('linkGrp') link_grp = etree.Element('linkGrp')
link_grp.set('corresp', '#'+xml_id) link_grp.set('corresp', '#'+xml_id)
link_grp.set('targFunc', 'head argument') link_grp.set('targFunc', 'head argument')
@ -67,6 +103,23 @@ class Sentence:
link.set('target', '#' + xml_id + '.' + link_ref + ' #' + xml_id + '.' + str(link_id + 1)) link.set('target', '#' + xml_id + '.' + link_ref + ' #' + xml_id + '.' + str(link_id + 1))
link_grp.append(link) link_grp.append(link)
base.append(link_grp) base.append(link_grp)
# srl linkGrp
if self.srl_links:
link_grp = etree.Element('linkGrp')
link_grp.set('corresp', '#' + xml_id)
link_grp.set('targFunc', 'head argument')
link_grp.set('type', 'SRL')
for link_id, item in enumerate(self.srl_links):
link_ref, link_type = item
link = etree.Element('link')
link.set('ana', 'srl:' + link_type.replace(':', '_'))
if link_ref == u'0':
link.set('target', '#' + xml_id + ' #' + xml_id + '.' + str(link_id + 1))
else:
link.set('target', '#' + xml_id + '.' + link_ref + ' #' + xml_id + '.' + str(link_id + 1))
link_grp.append(link)
base.append(link_grp)
return base return base
@ -93,35 +146,28 @@ class Paragraph:
class TeiDocument: class TeiDocument:
def __init__(self, _id, paragraphs=list()): def __init__(self, _id, paragraphs=list(), metadata=None):
self._id = _id self._id = _id
self.metadata = metadata
self.paragraphs = paragraphs self.paragraphs = paragraphs
def as_xml(self): def as_xml(self):
root = etree.Element('TEI') root = etree.Element('div')
root.set('xmlns', 'http://www.tei-c.org/ns/1.0')
set_xml_attr(root, 'lang', 'sl')
xml_id = self._id xml_id = self._id
if xml_id is not None: if xml_id is not None:
set_xml_attr(root, 'id', xml_id) set_xml_attr(root, 'id', xml_id)
tei_header = etree.SubElement(root, 'teiHeader') bibl = etree.Element('bibl')
bibl.set('corresp', f'#{xml_id}')
text = etree.SubElement(root, 'text') bibl.set('n', f'#{xml_id}')
body = etree.SubElement(text, 'body') for k, v in self.metadata.items():
bibl_el = etree.Element(k)
bibl_el.text = v
bibl.append(bibl_el)
root.append(bibl)
for para in self.paragraphs: for para in self.paragraphs:
body.append(para.as_xml(id_prefix=xml_id)) root.append(para.as_xml(id_prefix=xml_id))
encoding_desc = etree.SubElement(tei_header, 'encodingDesc')
tags_decl = etree.SubElement(encoding_desc, 'tagsDecl')
namespace = etree.SubElement(tags_decl, 'namespace')
namespace.set('name', 'http://www.tei-c.org/ns/1.0')
for tag in ['p', 's', 'pc', 'w']:
count = int(text.xpath('count(.//{})'.format(tag)))
tag_usage = etree.SubElement(namespace, 'tagUsage')
tag_usage.set('gi', tag)
tag_usage.set('occurs', str(count))
return root return root
def add_paragraph(self, paragraph): def add_paragraph(self, paragraph):
@ -129,10 +175,13 @@ class TeiDocument:
def build_tei_etrees(documents): def build_tei_etrees(documents):
elements = [] root = etree.Element('body')
root.set('xmlns', 'http://www.tei-c.org/ns/1.0')
set_xml_attr(root, 'base', 'korpus.xml')
set_xml_attr(root, 'lang', 'sl')
for document in documents: for document in documents:
elements.append(document.as_xml()) root.append(document.as_xml())
return elements return root
def set_xml_attr(node, attribute, value): def set_xml_attr(node, attribute, value):
@ -155,11 +204,12 @@ def is_metaline(line):
return False return False
def construct_tei_documents(conllu_lines): def construct_tei_documents(conllu_lines, metadata):
documents = [] documents = []
doc_id = None doc_id = None
document_paragraphs = [] doc_id_num = 0
document_paragraphs = []
para_id = None para_id = None
para_buffer = [] para_buffer = []
@ -171,9 +221,12 @@ def construct_tei_documents(conllu_lines):
if len(para_buffer) > 0: if len(para_buffer) > 0:
document_paragraphs.append(construct_paragraph(para_id, para_buffer)) document_paragraphs.append(construct_paragraph(para_id, para_buffer))
if len(document_paragraphs) > 0: if len(document_paragraphs) > 0:
print(metadata)
print(doc_id_num)
documents.append( documents.append(
TeiDocument(doc_id, document_paragraphs)) TeiDocument(doc_id, document_paragraphs, metadata[doc_id_num]))
document_paragraphs = [] document_paragraphs = []
doc_id_num += 1
doc_id = val doc_id = val
elif key == 'newpar id': elif key == 'newpar id':
if len(para_buffer) > 0: if len(para_buffer) > 0:
@ -191,7 +244,8 @@ def construct_tei_documents(conllu_lines):
if len(document_paragraphs) > 0: if len(document_paragraphs) > 0:
documents.append( documents.append(
TeiDocument(doc_id, document_paragraphs)) TeiDocument(doc_id, document_paragraphs, metadata[doc_id_num]))
doc_id_num += 1
return documents return documents
@ -234,7 +288,7 @@ def construct_sentence(sent_id, lines):
upos_other = tokens[5] upos_other = tokens[5]
depparse_link = tokens[6] depparse_link = tokens[6]
depparse_link_name = tokens[7] depparse_link_name = tokens[7]
misc = tokens[9] misc = {el.split('=')[0]: el.split('=')[1] for el in tokens[9].split('|')}
sentence.add_item( sentence.add_item(
token, token,
@ -247,17 +301,22 @@ def construct_sentence(sent_id, lines):
sentence.add_link( sentence.add_link(
depparse_link, depparse_link,
depparse_link_name) depparse_link_name)
if 'SRL' in misc:
sentence.add_srl_link(
depparse_link,
misc['SRL'])
return sentence return sentence
def construct_tei_etrees(conllu_lines): def construct_tei_etrees(conllu_lines, metadata):
documents = construct_tei_documents(conllu_lines) documents = construct_tei_documents(conllu_lines, metadata)
return build_tei_etrees(documents) return build_tei_etrees(documents)
def convert_file(input_file_name, output_file_name): def convert_file(input_file_name, output_file_name, metadata):
input_file = open(input_file_name, 'r') input_file = open(input_file_name, 'r')
root = construct_tei_etrees(input_file)[0] root = construct_tei_etrees(input_file, metadata)
tree = etree.ElementTree(root) tree = etree.ElementTree(root)
tree.write(output_file_name, encoding='UTF-8', pretty_print=True) tree.write(output_file_name, encoding='UTF-8', pretty_print=True)
input_file.close() input_file.close()

View File

@ -261,8 +261,8 @@ class Converter:
level information. level information.
""" """
if (msd.code not in self.specifications.codes_map[msd.language]): # if (msd.code not in self.specifications.codes_map[msd.language]):
raise ConverterException('The msd {} is unknown'.format(msd.code)) # raise ConverterException('The msd {} is unknown'.format(msd.code))
category_char = msd.code[0].lower() category_char = msd.code[0].lower()
value_chars = msd.code[1:] value_chars = msd.code[1:]

59
run.py Normal file
View File

@ -0,0 +1,59 @@
import os
from conversion_utils.conllu_to_tei import convert_file
import csv
# dir_path = 'data/conllu'
# out_dir_path = 'data/tei'
# for filename in os.listdir(dir_path):
# in_name = os.path.join(dir_path, filename)
# out_filename = filename.split('.')[:-1]
# out_filename = '.'.join(out_filename) + '.xml'
# out_name = os.path.join(out_dir_path, out_filename)
# convert_file(in_name, out_name)
metadata_list = []
with open('data/metadata.csv', newline='') as csvfile:
for line in csv.reader(csvfile):
metadata_list.append(line)
metadata = [{} for i in range(len(metadata_list[0]) - 1)]
for i in range(1, len(metadata_list[0])):
metadata[i - 1]['title'] = metadata_list[0][i]
metadata[i - 1]['subtitle'] = metadata_list[1][i]
metadata[i - 1]['authors'] = metadata_list[2][i]
metadata[i - 1]['first_edition'] = metadata_list[3][i]
metadata[i - 1]['edition_in_corpus'] = metadata_list[4][i]
metadata[i - 1]['layer_according_to_SEJO'] = metadata_list[5][i]
metadata[i - 1]['audience'] = metadata_list[6][i]
metadata[i - 1]['hours_of_classes'] = metadata_list[7][i]
metadata[i - 1]['publisher'] = metadata_list[8][i]
metadata[i - 1]['file_name'] = metadata_list[9][i]
dir_path = 'data/conllu'
out_path = 'data/tei/tei.xml'
out_dir = 'data/conllu.conllu'
# out_dir = dir_path + '/conllu_small.conllu'
metadata_indices = [6, 16, 13, 7, 2, 1, 3, 14, 15, 0, 8, 4, 11, 9, 12, 5, 10]
out_file = open(out_dir, 'w')
metadata_indices = []
for fn_i, filename in enumerate(os.listdir(dir_path)):
in_name = os.path.join(dir_path, filename)
out_filename = filename.split('.')[:-1]
out_filename = '.'.join(out_filename)
for m_i, el in enumerate(metadata):
if el['file_name'] == out_filename:
metadata_indices.append(m_i)
out_filename = out_filename + '.xml'
out_name = os.path.join(out_dir, out_filename)
in_file = open(in_name, 'r')
data = f'# newdoc id = doc{str(fn_i+1)}\n'
data += in_file.read()
in_file.close()
out_file.write(data)
out_file.close()
shuffled_metadata = [metadata[el] for el in metadata_indices]
convert_file(out_dir, out_path, shuffled_metadata)