Added new ssj500k reading option. Fixed GOS taxonomy
This commit is contained in:
@@ -12,6 +12,7 @@ import java.util.concurrent.atomic.AtomicLong;
|
||||
import data.CalculateFor;
|
||||
import data.Filter;
|
||||
import data.MultipleHMKeys;
|
||||
import gui.ValidationUtil;
|
||||
import org.apache.commons.csv.CSVFormat;
|
||||
import org.apache.commons.csv.CSVPrinter;
|
||||
import org.apache.commons.csv.QuoteMode;
|
||||
@@ -87,6 +88,9 @@ public class Export {
|
||||
|
||||
|
||||
//CSV file header
|
||||
if (!ValidationUtil.isEmpty(filter.getSkipValue()) && filter.getSkipValue() > 0) {
|
||||
FILE_HEADER_AL.add("Izpuščene besede");
|
||||
}
|
||||
FILE_HEADER_AL.add(filter.getCalculateFor().toHeaderString());
|
||||
if (filter.getCalculateFor().equals(CalculateFor.LEMMA))
|
||||
FILE_HEADER_AL.add("Lema male črke");
|
||||
@@ -125,11 +129,9 @@ public class Export {
|
||||
|
||||
// for (Map<MultipleHMKeys, AtomicLong> value : taxonomyResults.values()) {
|
||||
for (CalculateFor otherKey : filter.getMultipleKeys()) {
|
||||
if (num_taxonomy_frequencies.get(otherKey) > 0) {
|
||||
FILE_HEADER_AL.add(otherKey.toHeaderString());
|
||||
if (otherKey.equals(CalculateFor.LEMMA))
|
||||
FILE_HEADER_AL.add("Lema male črke");
|
||||
}
|
||||
FILE_HEADER_AL.add(otherKey.toHeaderString());
|
||||
if (otherKey.equals(CalculateFor.LEMMA))
|
||||
FILE_HEADER_AL.add("Lema male črke");
|
||||
}
|
||||
|
||||
// if(otherKey.equals(CalculateFor.LEMMA)){
|
||||
@@ -215,9 +217,12 @@ public class Export {
|
||||
|
||||
for (Map.Entry<MultipleHMKeys, Long> e : map.entrySet()) {
|
||||
List dataEntry = new ArrayList<>();
|
||||
dataEntry.add(e.getKey().getK1());
|
||||
if (!ValidationUtil.isEmpty(filter.getSkipValue()) && filter.getSkipValue() > 0) {
|
||||
dataEntry.add(e.getKey().getK1());
|
||||
}
|
||||
dataEntry.add(eraseSkipgramStars(e.getKey().getK1(), filter));
|
||||
if (filter.getCalculateFor().equals(CalculateFor.LEMMA)){
|
||||
dataEntry.add(e.getKey().getK1().toLowerCase());
|
||||
dataEntry.add(eraseSkipgramStars(e.getKey().getK1().toLowerCase(), filter));
|
||||
}
|
||||
|
||||
int i = 0;
|
||||
@@ -225,20 +230,20 @@ public class Export {
|
||||
switch(i){
|
||||
case 0:
|
||||
if (otherKey.equals(CalculateFor.LEMMA)){
|
||||
dataEntry.add(e.getKey().getK2());
|
||||
dataEntry.add(e.getKey().getK2().toLowerCase());
|
||||
dataEntry.add(eraseSkipgramStars(e.getKey().getK2(), filter));
|
||||
dataEntry.add(eraseSkipgramStars(e.getKey().getK2().toLowerCase(), filter));
|
||||
} else {
|
||||
dataEntry.add(e.getKey().getK2());
|
||||
dataEntry.add(eraseSkipgramStars(e.getKey().getK2(), filter));
|
||||
}
|
||||
break;
|
||||
case 1:
|
||||
dataEntry.add(e.getKey().getK3());
|
||||
dataEntry.add(eraseSkipgramStars(e.getKey().getK3(), filter));
|
||||
break;
|
||||
case 2:
|
||||
dataEntry.add(e.getKey().getK4());
|
||||
dataEntry.add(eraseSkipgramStars(e.getKey().getK4(), filter));
|
||||
break;
|
||||
case 3:
|
||||
dataEntry.add(e.getKey().getK5());
|
||||
dataEntry.add(eraseSkipgramStars(e.getKey().getK5(), filter));
|
||||
break;
|
||||
}
|
||||
|
||||
@@ -330,6 +335,13 @@ public class Export {
|
||||
return fileName;
|
||||
}
|
||||
|
||||
private static String eraseSkipgramStars(String s, Filter filter){
|
||||
if (!ValidationUtil.isEmpty(filter.getSkipValue()) && filter.getSkipValue() > 0) {
|
||||
s = s.replace("* ", "");
|
||||
}
|
||||
return s;
|
||||
}
|
||||
|
||||
public static String SetToCSV(String title, Object[][] result, File resultsPath, LinkedHashMap<String, String> headerInfoBlock) {
|
||||
//Delimiter used in CSV file
|
||||
String NEW_LINE_SEPARATOR = "\n";
|
||||
|
||||
Reference in New Issue
Block a user