You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
list/src/main/resources/message_en.properties

270 lines
13 KiB

# general
window.title=Corpus analyzer
hyperlink.help=Help
button.language=SL
button.computeNgrams=Calculate
button.cancel=Cancel
# template
tab.corpusTab=Corpus
tab.filterTab=Filter
tab.characterLevelTabNew=Characters
tab.wordLevelTab=Word parts
tab.oneWordAnalysisTab=Words
tab.stringLevelTabNew2=Word sets
# corpus tab
label.setCorpusLocation=Set corpus location
button.setCorpusLocation=Set location
label.readHeaderInfo=Read info from headers
checkBox.readHeaderInfo=
label.chooseResultsLocation=Choose result location
button.chooseResultsLocation=Set location
label.selectReader=Select reader
label.outputName=Output file name
label.corpusTab.chooseCorpusLocationH=Select the folder which contains the corpus. The folder should only contain one corpus and should not contain files that are not part of the corpus.
label.corpusTab.readHeaderInfoH=If you select this option, the taxonomy will be read separately. This might take a while.
label.corpusTab.chooseResultsLocationH=Choose result location
label.corpusTab.selectReaderH=Select reader
label.corpusTab.outputNameH=Output file name
# character analysis tab
label.stringLength=Number of characters
label.calculateFor=Calculate for
label.displayTaxonomy=Display taxonomies
label.dataLimit=Data limitations
label.msd=Morphosyntactic tag
label.taxonomy=Filter by taxonomy
label.minimalOccurrences=Min. nr. occurrences
label.minimalTaxonomy=Min. nr. tax. branches
label.taxonomySetOperation=Filtriraj taksonomijo po
label.solarFilters=Selected filters:
string.lemma=lemma
string.word=word
label.letter.stringLengthH=Enter the length of character strings.
label.letter.calculateForH=Character strings will be counted in the selected units.
label.letter.displayTaxonomyH=The output will also contain the distribution of character strings across the corpus taxonomy.
label.letter.msdH=Character strings will be counted only in words with the provided tag.
label.letter.taxonomyH=Character strings will be counted only in selected text types.
label.letter.minimalOccurrencesH=Character strings with fewer occurrences will not be included in the output.
label.letter.minimalTaxonomyH=Character strings that occur in fewer taxonomy branches will not be included in the output.
label.letter.taxonomySetOperationH=Izpisuj iz besedil, ki ustrezajo vsaj eni od izbranih vej (unija) ali vsem izbranim vejam (presek)
# word part tab
label.alsoVisualize=Also split by
label.lengthSearch=Search for word parts of a specified length
label.prefixLength=Length of initial part
label.suffixLength=Length of final part
label.listSearch=Search for word parts with a specified list
label.prefixList=List of initial parts
label.suffixList=List of final parts
label.wordPart.calculateForH=Word parts will be counted in the selected units.
label.wordPart.alsoVisualizeH=The output will also include the selected data.
label.wordPart.displayTaxonomyH=The output will also contain the distribution of word parts across the corpus taxonomy.
label.wordPart.prefixLengthH=Specify the length (in number of characters) of the initial word part.
label.wordPart.suffixLengthH=Specify the length (in number of characters) of the final word part.
label.wordPart.prefixListH=Separate the word parts with a semicolon (e.g. out; over)
label.wordPart.suffixListH=Separate the word parts with a semicolon (e.g. ation; ness).
label.wordPart.msdH=Word parts will only be counted in words with the specified tag.
label.wordPart.taxonomyH=Word parts will only be counted in the selected text types.
label.wordPart.minimalOccurrencesH=Units with the specified word part that occur fewer times will not be included in the output.
label.wordPart.minimalTaxonomyH=Units with the specified word part that are present in fewer taxonomy branches will not be included in the output.
# word tab
label.writeMsdAtTheEnd=Split the morphosyntactic tag
label.word.calculateForH=Specify what the program should treat as the main unit for the output.
label.word.alsoVisualizeH=The output will also contain the selected data.
label.word.displayTaxonomyH=The output will also contain the distribution of units across the corpus taxonomy.
label.word.writeMsdAtTheEndH=The output will also include individual parts of morphosyntactic tags.
label.word.msdH=Only words with the specified tag will be counted.
label.word.taxonomyH=Only words in the selected text types will be counted.
label.word.minimalOccurrencesH=Words with fewer occurrences will not be included in the output.
label.word.minimalTaxonomyH=Words that occur in fewer taxonomy branches will not be included in the output.
# word sets tab
label.wordSet.calculateForH=Specify the units from which word sets will be extracted.
label.wordSet.alsoVisualizeH=The output will also include the selected data.
label.wordSet.displayTaxonomyH=The output will also contain the distribution of word sets across the corpus taxonomy.
label.wordSet.skipValueH=Enter the maximum number of words that can appear between two words in a word set.
label.wordSet.ngramValueH=The program will extract word sets with the specified number of tokens.
label.wordSet.notePunctuationsH=Word sets will include punctuation.
label.wordSet.collocabilityH=The program will also calculate collocability measures between words within the word set.
label.wordSet.msdH=The program will only count word sets with the specified tag.
label.wordSet.taxonomyH=Word sets will only be extracted from the selected taxonomy branches.
label.wordSet.minimalOccurrencesH=Word sets with fewer occurrences will not be included in the output.
label.wordSet.minimalTaxonomyH=Word sets that occur in fewer taxonomy branches will not be included in the output.
# calculate for
calculateFor.WORD=word
calculateFor.NORMALIZED_WORD=normalized word
calculateFor.LEMMA=lemma
calculateFor.MORPHOSYNTACTIC_SPECS=morphosyntactic tag
calculateFor.MORPHOSYNTACTIC_PROPERTY=morphosyntactic property
calculateFor.WORD_TYPE=word type
calculateFor.DIST_WORDS=word
calculateFor.DIST_LEMMAS=lemma
# n-grams
label.skipValue=Skip value
label.slowSpeedWarning=WARNING! USING THE ABOVE FILTER MAY DECREASE PROCESSING SPEED!
label.ngramValue=N-gram length
label.notePunctuations=Include punctuation
label.collocability=Collocability
# taxonomy set operations
taxonomySetOperation.UNION=union
taxonomySetOperation.INTERSECTION=intersection
# filtersSolar
filter.solarRegijaL=Region
filter.solarPredmetL=Subject
filter.solarRazredL=Class
filter.solarLetoL=Year
filter.solarSolaL=School
filter.solarVrstaBesedilaL=Text type
filter.solarRegija=region
filter.solarPredmet=subject
filter.solarRazred=class
filter.solarLeto=year
filter.solarSola=school
filter.solarVrstaBesedila=type
# messages
message.WARNING_CORPUS_NOT_FOUND=No suitable corpus files have been found in the selected directory.
message.WARNING_RESULTS_DIR_NOT_VALID=You do not have permission to access the selected directory.
message.WARNING_DIFFERING_NGRAM_LEVEL_AND_FILTER_TOKENS=The specified n-gram length and number of words do not match.
message.WARNING_DIFFERING_NGRAM_LEVEL_AND_FILTER_TOKENS_INFO=Choose another number or modify the filter.
message.WARNING_WORD_OR_LEMMA=Specify if you want to calculate statistics for words or lemmas.
message.WARNING_ONLY_NUMBERS_ALLOWED=Please enter a valid number.
message.WARNING_NUMBER_TOO_BIG=The entered number is larger than the number of taxonomy branches.
message.WARNING_MISMATCHED_NGRAM_AND_TOKENS_VALUES=The number for n-grams (%d) and number of tags included (%d) must match.
message.WARNING_MISSING_STRING_LENGTH=String length must be higher than 0. Length is set to default value (1).
message.WARNING_NO_TAXONOMY_FOUND=The program was unable to read the taxonomy from the corpus files. Please select another directory or a different corpus.
message.WARNING_NO_SOLAR_FILTERS_FOUND=The program was unable to read the filters from corpus files. Please select another location or a different corpus.
message.ERROR_WHILE_EXECUTING=An error occurred during program execution.
message.ERROR_WHILE_SAVING_RESULTS_TO_CSV=An error occurred while saving results.
message.ERROR_NOT_ENOUGH_MEMORY=Your memory is insufficient for analyzing such a large amount of data.
message.ERROR_NO_REGI_FILE_FOUND=Missing file \"%s\".
message.MISSING_NGRAM_LEVEL=N-gram level
message.MISSING_CALCULATE_FOR=Calculate for
message.MISSING_SKIP=""
message.MISSING_STRING_LENGTH=String length
message.MISMATCHED_STRING_LENGTH_AND_MSD_REGEX=String length and regex filter do not match.
message.NOTIFICATION_FOUND_X_FILES=Nr. of found files: %s
message.NOTIFICATION_CORPUS=Corpus: %s
message.NOTIFICATION_ANALYSIS_COMPLETED=Analysis complete. The results have been saved successfully.
message.NOTIFICATION_ANALYSIS_COMPLETED_NO_RESULTS=Analysis complete, but it was not possible to calculate statistics to match all the specified conditions.
message.RESULTS_PATH_SET_TO_DEFAULT=Save location is set to corpus location.
message.NOTIFICATION_ANALYSIS_CANCELED=The analysis was canceled.
message.ONGOING_NOTIFICATION_ANALYZING_FILE_X_OF_Y=Analyzing file %d of %d (%s) - Estimated time remaining %d s
message.CANCELING_NOTIFICATION=Canceled
message.LABEL_CORPUS_LOCATION_NOT_SET=Corpus location is not set.
message.LABEL_RESULTS_LOCATION_NOT_SET=Result location is not set.
message.LABEL_RESULTS_CORPUS_TYPE_NOT_SET=Corpus type is not set.
message.LABEL_SCANNING_CORPUS=Searching for and analyzing corpus files...
message.LABEL_SCANNING_SINGLE_FILE_CORPUS=Input analysis
message.COMPLETED=Completed
#message.TOOLTIP_chooseCorpusLocationB=Select folder which contains corpus. The folder should only contain one corpus and should not contain too many files that are not part of corpus.
#message.TOOLTIP_readHeaderInfoChB=If you select this option taxonomy will be read separately. This might take a while.
message.TOOLTIP_readNotePunctuationsChB=The punctuation in sentences is included in the analysis.
message.TOOLTIP_readDisplayTaxonomyChB=The output file will include the distribution across the taxonomy branches.
windowTitles.error=Error
windowTitles.warning=Warning
windowTitles.confirmation=Confirmation
# export header translations
exportHeader.corpus=Corpus:
exportHeader.date=Date:
exportHeader.executionTime=Execution time:
exportHeader.analysis=Analysis:
exportHeader.analysis.letters=characters
exportHeader.analysis.wordParts=word parts
exportHeader.analysis.words=words
exportHeader.analysis.wordSets=word sets
exportHeader.numberLetters=Number of characters:
exportHeader.calculateFor=Calculate for:
exportHeader.alsoFilter=Also split by:
exportHeader.displayTaxonomies=Display taxonomy branches:
exportHeader.ngramLevel=N-gram level:
exportHeader.skipValue=Skip value:
exportHeader.notePunctuations=Include punctuation:
exportHeader.collocability=Collocability:
exportHeader.writeMSDAtTheEnd=Write tag at the end:
exportHeader.prefixLength=Initial part length:
exportHeader.suffixLength=Final part length:
exportHeader.prefixList=Initial part list:
exportHeader.suffixList=Final part list:
exportHeader.msd=Morphosyntactic tag:
exportHeader.taxonomy=Filter by taxonomy:
exportHeader.minOccurrences=Min. nr. occurrences:
exportHeader.minTaxonomies=Min. nr. taxonomy branches:
exportHeader.additionalFilters=Additional filters:
exportHeader.yes=yes
exportHeader.no=no
exportHeader.taxonomySetOperation=Filter taxonomy by:
# export table header translations
exportTable.skippedWords=Skipped words
exportTable.lettersSmall=Characters (lower case)
exportTable.wordsSmall=Lemma (lower case)
exportTable.wordBeginning=Initial part of the word
exportTable.wordEnding=Final part of the word
exportTable.wordRest=The rest of the word
exportTable.totalRelativeFrequency=Total relative frequency (over one million occurrences)
exportTable.absoluteFrequency=Absolute frequency
exportTable.percentage=Share
exportTable.relativeFrequency=Relative frequency
exportTable.msd=msd
# parts
exportTable.part.word=words:
exportTable.part.normalizedWord=normalized words:
exportTable.part.lemma=lemmas:
exportTable.part.msd=msd:
exportTable.part.msdProperty=msd property:
exportTable.part.wordType=word type:
exportTable.part.letterSet=character set
exportTable.part.word2=word
exportTable.part.normalizedWord2=normalized word
exportTable.part.lemma2=lemma
exportTable.part.msd2=msd
exportTable.part.msdProperty2=msd property
exportTable.part.wordType2=word type
exportTable.part.letterSet2=Share of total sum of all letter sets
exportTable.part.letterSet3=Letter set
exportTable.part.word3=Word
exportTable.part.normalizedWord3=Normalized word
exportTable.part.lemma3=Lemma
exportTable.part.msd3=Msd
exportTable.part.msdProperty3=Msd property
exportTable.part.wordType3=Word type
exportTable.part.set=set
exportTable.part.share=Absolute share of
exportTable.part.absoluteFrequency=Absolute frequency of
exportTable.part.totalFound=Total sum of all
exportTable.part.totalFoundLetters=Total sum of all found letters of
exportTable.part.totalSumString=Total sum of
exportTable.part.totalSumLetters=Total sum of all letters of
# generated files names
exportFileName.letters=letters
exportFileName.wordParts=word-parts
exportFileName.words=words
exportFileName.wordSets=word-sets
exportFileName.gram=-gram
exportFileName.skip=-skip