| predictLanguages(CharSequence) |  | 0% |  | 0% | 9 | 9 | 29 | 29 | 1 | 1 |
| seenEnough(LinkedList, Language[], Map) |  | 0% |  | 0% | 9 | 9 | 24 | 24 | 1 | 1 |
| predict(Map) |  | 0% |  | 0% | 3 | 3 | 14 | 14 | 1 | 1 |
| chunk(CharSequence, int, int) |  | 0% |  | 0% | 3 | 3 | 6 | 6 | 1 | 1 |
| getSupportedLanguages() |  | 0% |  | 0% | 2 | 2 | 5 | 5 | 1 | 1 |
| ProbingLanguageDetector(LanguageDetectorModel, CharSequenceNormalizer[]) |  | 0% | | n/a | 1 | 1 | 8 | 8 | 1 | 1 |
| setMinDiff(double) |  | 0% |  | 0% | 2 | 2 | 4 | 4 | 1 | 1 |
| predictLanguage(CharSequence) |  | 0% | | n/a | 1 | 1 | 1 | 1 | 1 | 1 |
| lambda$predict$0(Language, Language) |  | 0% | | n/a | 1 | 1 | 1 | 1 | 1 | 1 |
| setChunkSize(int) |  | 0% | | n/a | 1 | 1 | 2 | 2 | 1 | 1 |
| setMinConsecImprovements(int) |  | 0% | | n/a | 1 | 1 | 2 | 2 | 1 | 1 |
| setMaxLength(int) |  | 0% | | n/a | 1 | 1 | 2 | 2 | 1 | 1 |
| setNormalizer(CharSequenceNormalizer) |  | 0% | | n/a | 1 | 1 | 2 | 2 | 1 | 1 |
| getChunkSize() |  | 0% | | n/a | 1 | 1 | 1 | 1 | 1 | 1 |
| getMinConsecImprovements() |  | 0% | | n/a | 1 | 1 | 1 | 1 | 1 | 1 |
| getMinDiff() |  | 0% | | n/a | 1 | 1 | 1 | 1 | 1 | 1 |
| getMaxLength() |  | 0% | | n/a | 1 | 1 | 1 | 1 | 1 | 1 |