From d135cc7f2bd1524987eafeb38969d625f92ed2b9 Mon Sep 17 00:00:00 2001 From: voje Date: Thu, 11 Apr 2019 07:49:31 +0200 Subject: [PATCH] fixed the sentence duplication bug. todo: functors --- src/backend_flask/app.py | 3 +++ src/pkg/valency/valency/Frame.py | 2 +- src/pkg/valency/valency/reduce_functions.py | 10 ++++++---- 3 files changed, 10 insertions(+), 5 deletions(-) diff --git a/src/backend_flask/app.py b/src/backend_flask/app.py index 7127a72..fb05a2c 100644 --- a/src/backend_flask/app.py +++ b/src/backend_flask/app.py @@ -302,6 +302,9 @@ def api_get_functor_frames(): for ent in cur: frames += frames_from_db_entry(ent) # pre-process this step for prod TODO + for f in frames: + print(f.to_json()) + # raw_frames = vallex.functors_index[functor] # TODO ret_frames = RF(frames, valdb[SENSEMAP_COLL]) return prepare_frames(ret_frames) diff --git a/src/pkg/valency/valency/Frame.py b/src/pkg/valency/valency/Frame.py index 353ca5a..e02fe9f 100644 --- a/src/pkg/valency/valency/Frame.py +++ b/src/pkg/valency/valency/Frame.py @@ -30,7 +30,7 @@ def frames_from_db_entry(dbent): ], # sentences=[(dbent["sid"], dbent["tokens"])], sentences=[ - [(_full_tid(t["tid"]), t)for t in dbent["tokens"]], + [(_full_tid(t["tid"]), t) for t in dbent["tokens"]], ] )] return frames diff --git a/src/pkg/valency/valency/reduce_functions.py b/src/pkg/valency/valency/reduce_functions.py index 34853e2..50bd5e1 100644 --- a/src/pkg/valency/valency/reduce_functions.py +++ b/src/pkg/valency/valency/reduce_functions.py @@ -24,12 +24,14 @@ def sorted_by_len_tids(frames): def reduce_0(frames, valdb_sensemap=None): # new request... frames should be sorded by - # functors list (basically reduce_1, just each - # sentence gets its own frame) - r1_frames = reduce_1(frames) + # functors list + + # return frames # simplify code... + # frames = reduce_1(frames) + sorting_strings = [] separated_frames = [] - for frame in r1_frames: + for frame in frames: for tid in frame.tids: tmp_frame = DC(frame) tmp_frame.tids = [tid]