Coverage Report

Created: 2024-02-28 06:46

/src/tesseract/src/classify/intmatcher.cpp
Line
Count
Source (jump to first uncovered line)
1
/******************************************************************************
2
 ** Filename:    intmatcher.cpp
3
 ** Purpose:     Generic high level classification routines.
4
 ** Author:      Robert Moss
5
 ** (c) Copyright Hewlett-Packard Company, 1988.
6
 ** Licensed under the Apache License, Version 2.0 (the "License");
7
 ** you may not use this file except in compliance with the License.
8
 ** You may obtain a copy of the License at
9
 ** http://www.apache.org/licenses/LICENSE-2.0
10
 ** Unless required by applicable law or agreed to in writing, software
11
 ** distributed under the License is distributed on an "AS IS" BASIS,
12
 ** WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13
 ** See the License for the specific language governing permissions and
14
 ** limitations under the License.
15
 ******************************************************************************/
16
17
// Include automatically generated configuration file if running autoconf.
18
#ifdef HAVE_CONFIG_H
19
#  include "config_auto.h"
20
#endif
21
22
#include "intmatcher.h"
23
24
#include "classify.h"
25
#include "float2int.h"
26
#include "fontinfo.h"
27
#include "intproto.h"
28
#include "scrollview.h"
29
#include "shapetable.h"
30
31
#include "helpers.h"
32
33
#include <cassert>
34
#include <cmath>
35
36
namespace tesseract {
37
38
/*----------------------------------------------------------------------------
39
                    Global Data Definitions and Declarations
40
----------------------------------------------------------------------------*/
41
// Parameters of the sigmoid used to convert similarity to evidence in the
42
// similarity_evidence_table_ that is used to convert distance metric to an
43
// 8 bit evidence value in the secondary matcher. (See IntMatcher::Init).
44
const float IntegerMatcher::kSEExponentialMultiplier = 0.0f;
45
const float IntegerMatcher::kSimilarityCenter = 0.0075f;
46
47
static const uint8_t offset_table[] = {
48
    255, 0, 1, 0, 2, 0, 1, 0, 3, 0, 1, 0, 2, 0, 1, 0, 4, 0, 1, 0, 2, 0, 1, 0, 3, 0, 1, 0, 2,
49
    0,   1, 0, 5, 0, 1, 0, 2, 0, 1, 0, 3, 0, 1, 0, 2, 0, 1, 0, 4, 0, 1, 0, 2, 0, 1, 0, 3, 0,
50
    1,   0, 2, 0, 1, 0, 6, 0, 1, 0, 2, 0, 1, 0, 3, 0, 1, 0, 2, 0, 1, 0, 4, 0, 1, 0, 2, 0, 1,
51
    0,   3, 0, 1, 0, 2, 0, 1, 0, 5, 0, 1, 0, 2, 0, 1, 0, 3, 0, 1, 0, 2, 0, 1, 0, 4, 0, 1, 0,
52
    2,   0, 1, 0, 3, 0, 1, 0, 2, 0, 1, 0, 7, 0, 1, 0, 2, 0, 1, 0, 3, 0, 1, 0, 2, 0, 1, 0, 4,
53
    0,   1, 0, 2, 0, 1, 0, 3, 0, 1, 0, 2, 0, 1, 0, 5, 0, 1, 0, 2, 0, 1, 0, 3, 0, 1, 0, 2, 0,
54
    1,   0, 4, 0, 1, 0, 2, 0, 1, 0, 3, 0, 1, 0, 2, 0, 1, 0, 6, 0, 1, 0, 2, 0, 1, 0, 3, 0, 1,
55
    0,   2, 0, 1, 0, 4, 0, 1, 0, 2, 0, 1, 0, 3, 0, 1, 0, 2, 0, 1, 0, 5, 0, 1, 0, 2, 0, 1, 0,
56
    3,   0, 1, 0, 2, 0, 1, 0, 4, 0, 1, 0, 2, 0, 1, 0, 3, 0, 1, 0, 2, 0, 1, 0};
57
58
static const uint8_t next_table[] = {
59
    0,    0,    0,    0x2,  0,    0x4,  0x4,  0x6,  0,    0x8,  0x8,  0x0a, 0x08, 0x0c, 0x0c, 0x0e,
60
    0,    0x10, 0x10, 0x12, 0x10, 0x14, 0x14, 0x16, 0x10, 0x18, 0x18, 0x1a, 0x18, 0x1c, 0x1c, 0x1e,
61
    0,    0x20, 0x20, 0x22, 0x20, 0x24, 0x24, 0x26, 0x20, 0x28, 0x28, 0x2a, 0x28, 0x2c, 0x2c, 0x2e,
62
    0x20, 0x30, 0x30, 0x32, 0x30, 0x34, 0x34, 0x36, 0x30, 0x38, 0x38, 0x3a, 0x38, 0x3c, 0x3c, 0x3e,
63
    0,    0x40, 0x40, 0x42, 0x40, 0x44, 0x44, 0x46, 0x40, 0x48, 0x48, 0x4a, 0x48, 0x4c, 0x4c, 0x4e,
64
    0x40, 0x50, 0x50, 0x52, 0x50, 0x54, 0x54, 0x56, 0x50, 0x58, 0x58, 0x5a, 0x58, 0x5c, 0x5c, 0x5e,
65
    0x40, 0x60, 0x60, 0x62, 0x60, 0x64, 0x64, 0x66, 0x60, 0x68, 0x68, 0x6a, 0x68, 0x6c, 0x6c, 0x6e,
66
    0x60, 0x70, 0x70, 0x72, 0x70, 0x74, 0x74, 0x76, 0x70, 0x78, 0x78, 0x7a, 0x78, 0x7c, 0x7c, 0x7e,
67
    0,    0x80, 0x80, 0x82, 0x80, 0x84, 0x84, 0x86, 0x80, 0x88, 0x88, 0x8a, 0x88, 0x8c, 0x8c, 0x8e,
68
    0x80, 0x90, 0x90, 0x92, 0x90, 0x94, 0x94, 0x96, 0x90, 0x98, 0x98, 0x9a, 0x98, 0x9c, 0x9c, 0x9e,
69
    0x80, 0xa0, 0xa0, 0xa2, 0xa0, 0xa4, 0xa4, 0xa6, 0xa0, 0xa8, 0xa8, 0xaa, 0xa8, 0xac, 0xac, 0xae,
70
    0xa0, 0xb0, 0xb0, 0xb2, 0xb0, 0xb4, 0xb4, 0xb6, 0xb0, 0xb8, 0xb8, 0xba, 0xb8, 0xbc, 0xbc, 0xbe,
71
    0x80, 0xc0, 0xc0, 0xc2, 0xc0, 0xc4, 0xc4, 0xc6, 0xc0, 0xc8, 0xc8, 0xca, 0xc8, 0xcc, 0xcc, 0xce,
72
    0xc0, 0xd0, 0xd0, 0xd2, 0xd0, 0xd4, 0xd4, 0xd6, 0xd0, 0xd8, 0xd8, 0xda, 0xd8, 0xdc, 0xdc, 0xde,
73
    0xc0, 0xe0, 0xe0, 0xe2, 0xe0, 0xe4, 0xe4, 0xe6, 0xe0, 0xe8, 0xe8, 0xea, 0xe8, 0xec, 0xec, 0xee,
74
    0xe0, 0xf0, 0xf0, 0xf2, 0xf0, 0xf4, 0xf4, 0xf6, 0xf0, 0xf8, 0xf8, 0xfa, 0xf8, 0xfc, 0xfc, 0xfe};
75
76
// See http://b/19318793 (#6) for a complete discussion.
77
78
/**
79
 * Sort Key array in ascending order using heap sort
80
 * algorithm.  Also sort Index array that is tied to
81
 * the key array.
82
 * @param n Number of elements to sort
83
 * @param ra     Key array [1..n]
84
 * @param rb     Index array [1..n]
85
 */
86
3.26M
static void HeapSort(int n, int ra[], int rb[]) {
87
3.26M
  int i, rra, rrb;
88
3.26M
  int l, j, ir;
89
90
3.26M
  l = (n >> 1) + 1;
91
3.26M
  ir = n;
92
24.7M
  for (;;) {
93
24.7M
    if (l > 1) {
94
8.95M
      rra = ra[--l];
95
8.95M
      rrb = rb[l];
96
15.8M
    } else {
97
15.8M
      rra = ra[ir];
98
15.8M
      rrb = rb[ir];
99
15.8M
      ra[ir] = ra[1];
100
15.8M
      rb[ir] = rb[1];
101
15.8M
      if (--ir == 1) {
102
3.26M
        ra[1] = rra;
103
3.26M
        rb[1] = rrb;
104
3.26M
        return;
105
3.26M
      }
106
15.8M
    }
107
21.5M
    i = l;
108
21.5M
    j = l << 1;
109
54.8M
    while (j <= ir) {
110
33.3M
      if (j < ir && ra[j] < ra[j + 1]) {
111
10.2M
        ++j;
112
10.2M
      }
113
33.3M
      if (rra < ra[j]) {
114
25.8M
        ra[i] = ra[j];
115
25.8M
        rb[i] = rb[j];
116
25.8M
        j += (i = j);
117
25.8M
      } else {
118
7.53M
        j = ir + 1;
119
7.53M
      }
120
33.3M
    }
121
21.5M
    ra[i] = rra;
122
21.5M
    rb[i] = rrb;
123
21.5M
  }
124
3.26M
}
125
126
// Encapsulation of the intermediate data and computations made by the class
127
// pruner. The class pruner implements a simple linear classifier on binary
128
// features by heavily quantizing the feature space, and applying
129
// NUM_BITS_PER_CLASS (2)-bit weights to the features. Lack of resolution in
130
// weights is compensated by a non-constant bias that is dependent on the
131
// number of features present.
132
class ClassPruner {
133
public:
134
4.98M
  ClassPruner(int max_classes) {
135
    // The unrolled loop in ComputeScores means that the array sizes need to
136
    // be rounded up so that the array is big enough to accommodate the extra
137
    // entries accessed by the unrolling. Each pruner word is of sized
138
    // BITS_PER_WERD and each entry is NUM_BITS_PER_CLASS, so there are
139
    // BITS_PER_WERD / NUM_BITS_PER_CLASS entries.
140
    // See ComputeScores.
141
4.98M
    max_classes_ = max_classes;
142
4.98M
    rounded_classes_ =
143
4.98M
        RoundUp(max_classes, WERDS_PER_CP_VECTOR * BITS_PER_WERD / NUM_BITS_PER_CLASS);
144
4.98M
    class_count_ = new int[rounded_classes_];
145
4.98M
    norm_count_ = new int[rounded_classes_];
146
4.98M
    sort_key_ = new int[rounded_classes_ + 1];
147
4.98M
    sort_index_ = new int[rounded_classes_ + 1];
148
1.37G
    for (int i = 0; i < rounded_classes_; i++) {
149
1.36G
      class_count_[i] = 0;
150
1.36G
    }
151
4.98M
    pruning_threshold_ = 0;
152
4.98M
    num_features_ = 0;
153
4.98M
    num_classes_ = 0;
154
4.98M
  }
155
156
4.98M
  ~ClassPruner() {
157
4.98M
    delete[] class_count_;
158
4.98M
    delete[] norm_count_;
159
4.98M
    delete[] sort_key_;
160
4.98M
    delete[] sort_index_;
161
4.98M
  }
162
163
  /// Computes the scores for every class in the character set, by summing the
164
  /// weights for each feature and stores the sums internally in class_count_.
165
  void ComputeScores(const INT_TEMPLATES_STRUCT *int_templates, int num_features,
166
4.98M
                     const INT_FEATURE_STRUCT *features) {
167
4.98M
    num_features_ = num_features;
168
4.98M
    auto num_pruners = int_templates->NumClassPruners;
169
299M
    for (int f = 0; f < num_features; ++f) {
170
294M
      const INT_FEATURE_STRUCT *feature = &features[f];
171
      // Quantize the feature to NUM_CP_BUCKETS*NUM_CP_BUCKETS*NUM_CP_BUCKETS.
172
294M
      int x = feature->X * NUM_CP_BUCKETS >> 8;
173
294M
      int y = feature->Y * NUM_CP_BUCKETS >> 8;
174
294M
      int theta = feature->Theta * NUM_CP_BUCKETS >> 8;
175
294M
      int class_id = 0;
176
      // Each CLASS_PRUNER_STRUCT only covers CLASSES_PER_CP(32) classes, so
177
      // we need a collection of them, indexed by pruner_set.
178
2.66G
      for (unsigned pruner_set = 0; pruner_set < num_pruners; ++pruner_set) {
179
        // Look up quantized feature in a 3-D array, an array of weights for
180
        // each class.
181
2.37G
        const uint32_t *pruner_word_ptr = int_templates->ClassPruners[pruner_set]->p[x][y][theta];
182
7.11G
        for (int word = 0; word < WERDS_PER_CP_VECTOR; ++word) {
183
4.74G
          uint32_t pruner_word = *pruner_word_ptr++;
184
          // This inner loop is unrolled to speed up the ClassPruner.
185
          // Currently gcc would not unroll it unless it is set to O3
186
          // level of optimization or -funroll-loops is specified.
187
          /*
188
uint32_t class_mask = (1 << NUM_BITS_PER_CLASS) - 1;
189
for (int bit = 0; bit < BITS_PER_WERD/NUM_BITS_PER_CLASS; bit++) {
190
  class_count_[class_id++] += pruner_word & class_mask;
191
  pruner_word >>= NUM_BITS_PER_CLASS;
192
}
193
*/
194
4.74G
          class_count_[class_id++] += pruner_word & CLASS_PRUNER_CLASS_MASK;
195
4.74G
          pruner_word >>= NUM_BITS_PER_CLASS;
196
4.74G
          class_count_[class_id++] += pruner_word & CLASS_PRUNER_CLASS_MASK;
197
4.74G
          pruner_word >>= NUM_BITS_PER_CLASS;
198
4.74G
          class_count_[class_id++] += pruner_word & CLASS_PRUNER_CLASS_MASK;
199
4.74G
          pruner_word >>= NUM_BITS_PER_CLASS;
200
4.74G
          class_count_[class_id++] += pruner_word & CLASS_PRUNER_CLASS_MASK;
201
4.74G
          pruner_word >>= NUM_BITS_PER_CLASS;
202
4.74G
          class_count_[class_id++] += pruner_word & CLASS_PRUNER_CLASS_MASK;
203
4.74G
          pruner_word >>= NUM_BITS_PER_CLASS;
204
4.74G
          class_count_[class_id++] += pruner_word & CLASS_PRUNER_CLASS_MASK;
205
4.74G
          pruner_word >>= NUM_BITS_PER_CLASS;
206
4.74G
          class_count_[class_id++] += pruner_word & CLASS_PRUNER_CLASS_MASK;
207
4.74G
          pruner_word >>= NUM_BITS_PER_CLASS;
208
4.74G
          class_count_[class_id++] += pruner_word & CLASS_PRUNER_CLASS_MASK;
209
4.74G
          pruner_word >>= NUM_BITS_PER_CLASS;
210
4.74G
          class_count_[class_id++] += pruner_word & CLASS_PRUNER_CLASS_MASK;
211
4.74G
          pruner_word >>= NUM_BITS_PER_CLASS;
212
4.74G
          class_count_[class_id++] += pruner_word & CLASS_PRUNER_CLASS_MASK;
213
4.74G
          pruner_word >>= NUM_BITS_PER_CLASS;
214
4.74G
          class_count_[class_id++] += pruner_word & CLASS_PRUNER_CLASS_MASK;
215
4.74G
          pruner_word >>= NUM_BITS_PER_CLASS;
216
4.74G
          class_count_[class_id++] += pruner_word & CLASS_PRUNER_CLASS_MASK;
217
4.74G
          pruner_word >>= NUM_BITS_PER_CLASS;
218
4.74G
          class_count_[class_id++] += pruner_word & CLASS_PRUNER_CLASS_MASK;
219
4.74G
          pruner_word >>= NUM_BITS_PER_CLASS;
220
4.74G
          class_count_[class_id++] += pruner_word & CLASS_PRUNER_CLASS_MASK;
221
4.74G
          pruner_word >>= NUM_BITS_PER_CLASS;
222
4.74G
          class_count_[class_id++] += pruner_word & CLASS_PRUNER_CLASS_MASK;
223
4.74G
          pruner_word >>= NUM_BITS_PER_CLASS;
224
4.74G
          class_count_[class_id++] += pruner_word & CLASS_PRUNER_CLASS_MASK;
225
4.74G
        }
226
2.37G
      }
227
294M
    }
228
4.98M
  }
229
230
  /// Adjusts the scores according to the number of expected features. Used
231
  /// in lieu of a constant bias, this penalizes classes that expect more
232
  /// features than there are present. Thus an actual c will score higher for c
233
  /// than e, even though almost all the features match e as well as c, because
234
  /// e expects more features to be present.
235
4.98M
  void AdjustForExpectedNumFeatures(const uint16_t *expected_num_features, int cutoff_strength) {
236
1.32G
    for (int class_id = 0; class_id < max_classes_; ++class_id) {
237
1.32G
      if (num_features_ < expected_num_features[class_id]) {
238
235M
        int deficit = expected_num_features[class_id] - num_features_;
239
235M
        class_count_[class_id] -=
240
235M
            class_count_[class_id] * deficit / (num_features_ * cutoff_strength + deficit);
241
235M
      }
242
1.32G
    }
243
4.98M
  }
244
245
  /// Zeros the scores for classes disabled in the unicharset.
246
  /// Implements the black-list to recognize a subset of the character set.
247
0
  void DisableDisabledClasses(const UNICHARSET &unicharset) {
248
0
    for (int class_id = 0; class_id < max_classes_; ++class_id) {
249
0
      if (!unicharset.get_enabled(class_id)) {
250
0
        class_count_[class_id] = 0; // This char is disabled!
251
0
      }
252
0
    }
253
0
  }
254
255
  /** Zeros the scores of fragments. */
256
0
  void DisableFragments(const UNICHARSET &unicharset) {
257
0
    for (int class_id = 0; class_id < max_classes_; ++class_id) {
258
      // Do not include character fragments in the class pruner
259
      // results if disable_character_fragments is true.
260
0
      if (unicharset.get_fragment(class_id)) {
261
0
        class_count_[class_id] = 0;
262
0
      }
263
0
    }
264
0
  }
265
266
  /// Normalizes the counts for xheight, putting the normalized result in
267
  /// norm_count_. Applies a simple subtractive penalty for incorrect vertical
268
  /// position provided by the normalization_factors array, indexed by
269
  /// character class, and scaled by the norm_multiplier.
270
4.98M
  void NormalizeForXheight(int norm_multiplier, const uint8_t *normalization_factors) {
271
1.32G
    for (int class_id = 0; class_id < max_classes_; class_id++) {
272
1.32G
      norm_count_[class_id] =
273
1.32G
          class_count_[class_id] - ((norm_multiplier * normalization_factors[class_id]) >> 8);
274
1.32G
    }
275
4.98M
  }
276
277
  /** The nop normalization copies the class_count_ array to norm_count_. */
278
0
  void NoNormalization() {
279
0
    for (int class_id = 0; class_id < max_classes_; class_id++) {
280
0
      norm_count_[class_id] = class_count_[class_id];
281
0
    }
282
0
  }
283
284
  /// Prunes the classes using &lt;the maximum count> * pruning_factor/256 as a
285
  /// threshold for keeping classes. If max_of_non_fragments, then ignore
286
  /// fragments in computing the maximum count.
287
  void PruneAndSort(int pruning_factor, int keep_this, bool max_of_non_fragments,
288
4.98M
                    const UNICHARSET &unicharset) {
289
4.98M
    int max_count = 0;
290
1.32G
    for (int c = 0; c < max_classes_; ++c) {
291
1.32G
      if (norm_count_[c] > max_count &&
292
          // This additional check is added in order to ensure that
293
          // the classifier will return at least one non-fragmented
294
          // character match.
295
          // TODO(daria): verify that this helps accuracy and does not
296
          // hurt performance.
297
1.32G
          (!max_of_non_fragments || !unicharset.get_fragment(c))) {
298
18.8M
        max_count = norm_count_[c];
299
18.8M
      }
300
1.32G
    }
301
    // Prune Classes.
302
4.98M
    pruning_threshold_ = (max_count * pruning_factor) >> 8;
303
    // Select Classes.
304
4.98M
    if (pruning_threshold_ < 1) {
305
191k
      pruning_threshold_ = 1;
306
191k
    }
307
4.98M
    num_classes_ = 0;
308
1.32G
    for (int class_id = 0; class_id < max_classes_; class_id++) {
309
1.32G
      if (norm_count_[class_id] >= pruning_threshold_ || class_id == keep_this) {
310
20.6M
        ++num_classes_;
311
20.6M
        sort_index_[num_classes_] = class_id;
312
20.6M
        sort_key_[num_classes_] = norm_count_[class_id];
313
20.6M
      }
314
1.32G
    }
315
316
    // Sort Classes using Heapsort Algorithm.
317
4.98M
    if (num_classes_ > 1) {
318
3.26M
      HeapSort(num_classes_, sort_key_, sort_index_);
319
3.26M
    }
320
4.98M
  }
321
322
  /** Prints debug info on the class pruner matches for the pruned classes only.
323
   */
324
  void DebugMatch(const Classify &classify, const INT_TEMPLATES_STRUCT *int_templates,
325
0
                  const INT_FEATURE_STRUCT *features) const {
326
0
    int num_pruners = int_templates->NumClassPruners;
327
0
    int max_num_classes = int_templates->NumClasses;
328
0
    for (int f = 0; f < num_features_; ++f) {
329
0
      const INT_FEATURE_STRUCT *feature = &features[f];
330
0
      tprintf("F=%3d(%d,%d,%d),", f, feature->X, feature->Y, feature->Theta);
331
      // Quantize the feature to NUM_CP_BUCKETS*NUM_CP_BUCKETS*NUM_CP_BUCKETS.
332
0
      int x = feature->X * NUM_CP_BUCKETS >> 8;
333
0
      int y = feature->Y * NUM_CP_BUCKETS >> 8;
334
0
      int theta = feature->Theta * NUM_CP_BUCKETS >> 8;
335
0
      int class_id = 0;
336
0
      for (int pruner_set = 0; pruner_set < num_pruners; ++pruner_set) {
337
        // Look up quantized feature in a 3-D array, an array of weights for
338
        // each class.
339
0
        const uint32_t *pruner_word_ptr = int_templates->ClassPruners[pruner_set]->p[x][y][theta];
340
0
        for (int word = 0; word < WERDS_PER_CP_VECTOR; ++word) {
341
0
          uint32_t pruner_word = *pruner_word_ptr++;
342
0
          for (int word_class = 0; word_class < 16 && class_id < max_num_classes;
343
0
               ++word_class, ++class_id) {
344
0
            if (norm_count_[class_id] >= pruning_threshold_) {
345
0
              tprintf(" %s=%d,", classify.ClassIDToDebugStr(int_templates, class_id, 0).c_str(),
346
0
                      pruner_word & CLASS_PRUNER_CLASS_MASK);
347
0
            }
348
0
            pruner_word >>= NUM_BITS_PER_CLASS;
349
0
          }
350
0
        }
351
0
        tprintf("\n");
352
0
      }
353
0
    }
354
0
  }
355
356
  /** Prints a summary of the pruner result. */
357
  void SummarizeResult(const Classify &classify, const INT_TEMPLATES_STRUCT *int_templates,
358
                       const uint16_t *expected_num_features, int norm_multiplier,
359
0
                       const uint8_t *normalization_factors) const {
360
0
    tprintf("CP:%d classes, %d features:\n", num_classes_, num_features_);
361
0
    for (int i = 0; i < num_classes_; ++i) {
362
0
      int class_id = sort_index_[num_classes_ - i];
363
0
      std::string class_string = classify.ClassIDToDebugStr(int_templates, class_id, 0);
364
0
      tprintf(
365
0
          "%s:Initial=%d, E=%d, Xht-adj=%d, N=%d, Rat=%.2f\n", class_string.c_str(),
366
0
          class_count_[class_id], expected_num_features[class_id],
367
0
          (norm_multiplier * normalization_factors[class_id]) >> 8, sort_key_[num_classes_ - i],
368
0
          100.0 - 100.0 * sort_key_[num_classes_ - i] / (CLASS_PRUNER_CLASS_MASK * num_features_));
369
0
    }
370
0
  }
371
372
  /// Copies the pruned, sorted classes into the output results and returns
373
  /// the number of classes.
374
4.98M
  int SetupResults(std::vector<CP_RESULT_STRUCT> *results) const {
375
4.98M
    results->clear();
376
4.98M
    results->resize(num_classes_);
377
25.6M
    for (int c = 0; c < num_classes_; ++c) {
378
20.6M
      (*results)[c].Class = sort_index_[num_classes_ - c];
379
20.6M
      (*results)[c].Rating =
380
20.6M
          1.0f - sort_key_[num_classes_ - c] /
381
20.6M
                     (static_cast<float>(CLASS_PRUNER_CLASS_MASK) * num_features_);
382
20.6M
    }
383
4.98M
    return num_classes_;
384
4.98M
  }
385
386
private:
387
  /** Array[rounded_classes_] of initial counts for each class. */
388
  int *class_count_;
389
  /// Array[rounded_classes_] of modified counts for each class after
390
  /// normalizing for expected number of features, disabled classes, fragments,
391
  /// and xheights.
392
  int *norm_count_;
393
  /** Array[rounded_classes_ +1] of pruned counts that gets sorted */
394
  int *sort_key_;
395
  /** Array[rounded_classes_ +1] of classes corresponding to sort_key_. */
396
  int *sort_index_;
397
  /** Number of classes in this class pruner. */
398
  int max_classes_;
399
  /** Rounded up number of classes used for array sizes. */
400
  int rounded_classes_;
401
  /** Threshold count applied to prune classes. */
402
  int pruning_threshold_;
403
  /** The number of features used to compute the scores. */
404
  int num_features_;
405
  /** Final number of pruned classes. */
406
  int num_classes_;
407
};
408
409
/*----------------------------------------------------------------------------
410
              Public Code
411
----------------------------------------------------------------------------*/
412
/**
413
 * Runs the class pruner from int_templates on the given features, returning
414
 * the number of classes output in results.
415
 * @param int_templates          Class pruner tables
416
 * @param num_features           Number of features in blob
417
 * @param features               Array of features
418
 * @param normalization_factors  Array of fudge factors from blob
419
 *                               normalization process (by CLASS_INDEX)
420
 * @param expected_num_features  Array of expected number of features
421
 *                               for each class (by CLASS_INDEX)
422
 * @param results                Sorted Array of pruned classes. Must be an
423
 *                               array of size at least
424
 *                               int_templates->NumClasses.
425
 * @param keep_this
426
 */
427
int Classify::PruneClasses(const INT_TEMPLATES_STRUCT *int_templates, int num_features,
428
                           int keep_this, const INT_FEATURE_STRUCT *features,
429
                           const uint8_t *normalization_factors,
430
                           const uint16_t *expected_num_features,
431
4.98M
                           std::vector<CP_RESULT_STRUCT> *results) {
432
4.98M
  ClassPruner pruner(int_templates->NumClasses);
433
  // Compute initial match scores for all classes.
434
4.98M
  pruner.ComputeScores(int_templates, num_features, features);
435
  // Adjust match scores for number of expected features.
436
4.98M
  pruner.AdjustForExpectedNumFeatures(expected_num_features, classify_cp_cutoff_strength);
437
  // Apply disabled classes in unicharset - only works without a shape_table.
438
4.98M
  if (shape_table_ == nullptr) {
439
0
    pruner.DisableDisabledClasses(unicharset);
440
0
  }
441
  // If fragments are disabled, remove them, also only without a shape table.
442
4.98M
  if (disable_character_fragments && shape_table_ == nullptr) {
443
0
    pruner.DisableFragments(unicharset);
444
0
  }
445
446
  // If we have good x-heights, apply the given normalization factors.
447
4.98M
  if (normalization_factors != nullptr) {
448
4.98M
    pruner.NormalizeForXheight(classify_class_pruner_multiplier, normalization_factors);
449
4.98M
  } else {
450
0
    pruner.NoNormalization();
451
0
  }
452
  // Do the actual pruning and sort the short-list.
453
4.98M
  pruner.PruneAndSort(classify_class_pruner_threshold, keep_this, shape_table_ == nullptr,
454
4.98M
                      unicharset);
455
456
4.98M
  if (classify_debug_level > 2) {
457
0
    pruner.DebugMatch(*this, int_templates, features);
458
0
  }
459
4.98M
  if (classify_debug_level > 1) {
460
0
    pruner.SummarizeResult(*this, int_templates, expected_num_features,
461
0
                           classify_class_pruner_multiplier, normalization_factors);
462
0
  }
463
  // Convert to the expected output format.
464
4.98M
  return pruner.SetupResults(results);
465
4.98M
}
466
467
/**
468
 * IntegerMatcher returns the best configuration and rating
469
 * for a single class.  The class matched against is determined
470
 * by the uniqueness of the ClassTemplate parameter.  The
471
 * best rating and its associated configuration are returned.
472
 *
473
 * Globals:
474
 * - local_matcher_multiplier_ Normalization factor multiplier
475
 * param ClassTemplate Prototypes & tables for a class
476
 * param NumFeatures Number of features in blob
477
 * param Features Array of features
478
 * param NormalizationFactor Fudge factor from blob normalization process
479
 * param Result Class rating & configuration: (0.0 -> 1.0), 0=bad, 1=good
480
 * param Debug Debugger flag: 1=debugger on
481
 */
482
void IntegerMatcher::Match(INT_CLASS_STRUCT *ClassTemplate, BIT_VECTOR ProtoMask, BIT_VECTOR ConfigMask,
483
                           int16_t NumFeatures, const INT_FEATURE_STRUCT *Features,
484
                           UnicharRating *Result, int AdaptFeatureThreshold, int Debug,
485
20.6M
                           bool SeparateDebugWindows) {
486
20.6M
  auto *tables = new ScratchEvidence();
487
20.6M
  int Feature;
488
489
20.6M
  if (MatchDebuggingOn(Debug)) {
490
0
    tprintf("Integer Matcher -------------------------------------------\n");
491
0
  }
492
493
20.6M
  tables->Clear(ClassTemplate);
494
20.6M
  Result->feature_misses = 0;
495
496
1.39G
  for (Feature = 0; Feature < NumFeatures; Feature++) {
497
1.37G
    int csum = UpdateTablesForFeature(ClassTemplate, ProtoMask, ConfigMask, Feature,
498
1.37G
                                      &Features[Feature], tables, Debug);
499
    // Count features that were missed over all configs.
500
1.37G
    if (csum == 0) {
501
181M
      ++Result->feature_misses;
502
181M
    }
503
1.37G
  }
504
505
#ifndef GRAPHICS_DISABLED
506
  if (PrintProtoMatchesOn(Debug) || PrintMatchSummaryOn(Debug)) {
507
    DebugFeatureProtoError(ClassTemplate, ProtoMask, ConfigMask, *tables, NumFeatures, Debug);
508
  }
509
510
  if (DisplayProtoMatchesOn(Debug)) {
511
    DisplayProtoDebugInfo(ClassTemplate, ConfigMask, *tables, SeparateDebugWindows);
512
  }
513
514
  if (DisplayFeatureMatchesOn(Debug)) {
515
    DisplayFeatureDebugInfo(ClassTemplate, ProtoMask, ConfigMask, NumFeatures, Features,
516
                            AdaptFeatureThreshold, Debug, SeparateDebugWindows);
517
  }
518
#endif
519
520
20.6M
  tables->UpdateSumOfProtoEvidences(ClassTemplate, ConfigMask);
521
20.6M
  tables->NormalizeSums(ClassTemplate, NumFeatures);
522
523
20.6M
  FindBestMatch(ClassTemplate, *tables, Result);
524
525
#ifndef GRAPHICS_DISABLED
526
  if (PrintMatchSummaryOn(Debug)) {
527
    Result->Print();
528
  }
529
530
  if (MatchDebuggingOn(Debug)) {
531
    tprintf("Match Complete --------------------------------------------\n");
532
  }
533
#endif
534
535
20.6M
  delete tables;
536
20.6M
}
537
538
/**
539
 * FindGoodProtos finds all protos whose normalized proto-evidence
540
 * exceed AdaptProtoThreshold.  The list is ordered by increasing
541
 * proto id number.
542
 *
543
 * Globals:
544
 * - local_matcher_multiplier_    Normalization factor multiplier
545
 * param ClassTemplate Prototypes & tables for a class
546
 * param ProtoMask AND Mask for proto word
547
 * param ConfigMask AND Mask for config word
548
 * param NumFeatures Number of features in blob
549
 * param Features Array of features
550
 * param ProtoArray Array of good protos
551
 * param AdaptProtoThreshold Threshold for good protos
552
 * param Debug Debugger flag: 1=debugger on
553
 * @return Number of good protos in ProtoArray.
554
 */
555
int IntegerMatcher::FindGoodProtos(INT_CLASS_STRUCT *ClassTemplate, BIT_VECTOR ProtoMask,
556
                                   BIT_VECTOR ConfigMask, int16_t NumFeatures,
557
                                   INT_FEATURE_ARRAY Features, PROTO_ID *ProtoArray,
558
425
                                   int AdaptProtoThreshold, int Debug) {
559
425
  auto *tables = new ScratchEvidence();
560
425
  int NumGoodProtos = 0;
561
562
  /* DEBUG opening heading */
563
425
  if (MatchDebuggingOn(Debug)) {
564
0
    tprintf("Find Good Protos -------------------------------------------\n");
565
0
  }
566
567
425
  tables->Clear(ClassTemplate);
568
569
16.2k
  for (int Feature = 0; Feature < NumFeatures; Feature++) {
570
15.8k
    UpdateTablesForFeature(ClassTemplate, ProtoMask, ConfigMask, Feature, &(Features[Feature]),
571
15.8k
                           tables, Debug);
572
15.8k
  }
573
574
#ifndef GRAPHICS_DISABLED
575
  if (PrintProtoMatchesOn(Debug) || PrintMatchSummaryOn(Debug)) {
576
    DebugFeatureProtoError(ClassTemplate, ProtoMask, ConfigMask, *tables, NumFeatures, Debug);
577
  }
578
#endif
579
580
  /* Average Proto Evidences & Find Good Protos */
581
25.7k
  for (int proto = 0; proto < ClassTemplate->NumProtos; proto++) {
582
    /* Compute Average for Actual Proto */
583
25.3k
    int Temp = 0;
584
69.0k
    for (uint8_t i = 0; i < MAX_PROTO_INDEX && i < ClassTemplate->ProtoLengths[proto]; i++) {
585
43.6k
      Temp += tables->proto_evidence_[proto][i];
586
43.6k
    }
587
588
25.3k
    Temp /= ClassTemplate->ProtoLengths[proto];
589
590
    /* Find Good Protos */
591
25.3k
    if (Temp >= AdaptProtoThreshold) {
592
5.53k
      *ProtoArray = proto;
593
5.53k
      ProtoArray++;
594
5.53k
      NumGoodProtos++;
595
5.53k
    }
596
25.3k
  }
597
598
425
  if (MatchDebuggingOn(Debug)) {
599
0
    tprintf("Match Complete --------------------------------------------\n");
600
0
  }
601
425
  delete tables;
602
603
425
  return NumGoodProtos;
604
425
}
605
606
/**
607
 * FindBadFeatures finds all features with maximum feature-evidence <
608
 * AdaptFeatureThresh. The list is ordered by increasing feature number.
609
 * @param ClassTemplate Prototypes & tables for a class
610
 * @param ProtoMask AND Mask for proto word
611
 * @param ConfigMask AND Mask for config word
612
 * @param NumFeatures Number of features in blob
613
 * @param Features Array of features
614
 * @param FeatureArray Array of bad features
615
 * @param AdaptFeatureThreshold Threshold for bad features
616
 * @param Debug Debugger flag: 1=debugger on
617
 * @return Number of bad features in FeatureArray.
618
 */
619
int IntegerMatcher::FindBadFeatures(INT_CLASS_STRUCT *ClassTemplate, BIT_VECTOR ProtoMask,
620
                                    BIT_VECTOR ConfigMask, int16_t NumFeatures,
621
                                    INT_FEATURE_ARRAY Features, FEATURE_ID *FeatureArray,
622
425
                                    int AdaptFeatureThreshold, int Debug) {
623
425
  auto *tables = new ScratchEvidence();
624
425
  int NumBadFeatures = 0;
625
626
  /* DEBUG opening heading */
627
425
  if (MatchDebuggingOn(Debug)) {
628
0
    tprintf("Find Bad Features -------------------------------------------\n");
629
0
  }
630
631
425
  tables->Clear(ClassTemplate);
632
633
16.2k
  for (int Feature = 0; Feature < NumFeatures; Feature++) {
634
15.8k
    UpdateTablesForFeature(ClassTemplate, ProtoMask, ConfigMask, Feature, &Features[Feature],
635
15.8k
                           tables, Debug);
636
637
    /* Find Best Evidence for Current Feature */
638
15.8k
    int best = 0;
639
15.8k
    assert(ClassTemplate->NumConfigs < MAX_NUM_CONFIGS);
640
194k
    for (int i = 0; i < MAX_NUM_CONFIGS && i < ClassTemplate->NumConfigs; i++) {
641
178k
      if (tables->feature_evidence_[i] > best) {
642
12.6k
        best = tables->feature_evidence_[i];
643
12.6k
      }
644
178k
    }
645
646
    /* Find Bad Features */
647
15.8k
    if (best < AdaptFeatureThreshold) {
648
6.60k
      *FeatureArray = Feature;
649
6.60k
      FeatureArray++;
650
6.60k
      NumBadFeatures++;
651
6.60k
    }
652
15.8k
  }
653
654
#ifndef GRAPHICS_DISABLED
655
  if (PrintProtoMatchesOn(Debug) || PrintMatchSummaryOn(Debug)) {
656
    DebugFeatureProtoError(ClassTemplate, ProtoMask, ConfigMask, *tables, NumFeatures, Debug);
657
  }
658
#endif
659
660
425
  if (MatchDebuggingOn(Debug)) {
661
0
    tprintf("Match Complete --------------------------------------------\n");
662
0
  }
663
664
425
  delete tables;
665
425
  return NumBadFeatures;
666
425
}
667
668
IntegerMatcher::IntegerMatcher(tesseract::IntParam *classify_debug_level)
669
4
    : classify_debug_level_(classify_debug_level) {
670
  /* Initialize table for evidence to similarity lookup */
671
2.05k
  for (int i = 0; i < SE_TABLE_SIZE; i++) {
672
2.04k
    uint32_t IntSimilarity = i << (27 - SE_TABLE_BITS);
673
2.04k
    double Similarity = (static_cast<double>(IntSimilarity)) / 65536.0 / 65536.0;
674
2.04k
    double evidence = Similarity / kSimilarityCenter;
675
2.04k
    evidence = 255.0 / (evidence * evidence + 1.0);
676
677
2.04k
    if (kSEExponentialMultiplier > 0.0) {
678
0
      double scale =
679
0
          1.0 - std::exp(-kSEExponentialMultiplier) *
680
0
                    exp(kSEExponentialMultiplier * (static_cast<double>(i) / SE_TABLE_SIZE));
681
0
      evidence *= ClipToRange(scale, 0.0, 1.0);
682
0
    }
683
684
2.04k
    similarity_evidence_table_[i] = static_cast<uint8_t>(evidence + 0.5);
685
2.04k
  }
686
687
  /* Initialize evidence computation variables */
688
4
  evidence_table_mask_ = ((1 << kEvidenceTableBits) - 1) << (9 - kEvidenceTableBits);
689
4
  mult_trunc_shift_bits_ = (14 - kIntEvidenceTruncBits);
690
4
  table_trunc_shift_bits_ = (27 - SE_TABLE_BITS - (mult_trunc_shift_bits_ << 1));
691
4
  evidence_mult_mask_ = ((1 << kIntEvidenceTruncBits) - 1);
692
4
}
693
694
/*----------------------------------------------------------------------------
695
              Private Code
696
----------------------------------------------------------------------------*/
697
20.6M
void ScratchEvidence::Clear(const INT_CLASS_STRUCT *class_template) {
698
20.6M
  memset(sum_feature_evidence_, 0, class_template->NumConfigs * sizeof(sum_feature_evidence_[0]));
699
20.6M
  memset(proto_evidence_, 0, class_template->NumProtos * sizeof(proto_evidence_[0]));
700
20.6M
}
701
702
1.37G
void ScratchEvidence::ClearFeatureEvidence(const INT_CLASS_STRUCT *class_template) {
703
1.37G
  memset(feature_evidence_, 0, class_template->NumConfigs * sizeof(feature_evidence_[0]));
704
1.37G
}
705
706
/**
707
 * Print debugging information for Configurations
708
 */
709
static void IMDebugConfiguration(int FeatureNum, uint16_t ActualProtoNum, uint8_t Evidence,
710
0
                                 uint32_t ConfigWord) {
711
0
  tprintf("F = %3d, P = %3d, E = %3d, Configs = ", FeatureNum, static_cast<int>(ActualProtoNum),
712
0
          static_cast<int>(Evidence));
713
0
  while (ConfigWord) {
714
0
    if (ConfigWord & 1) {
715
0
      tprintf("1");
716
0
    } else {
717
0
      tprintf("0");
718
0
    }
719
0
    ConfigWord >>= 1;
720
0
  }
721
0
  tprintf("\n");
722
0
}
723
724
/**
725
 * Print debugging information for Configurations
726
 */
727
0
static void IMDebugConfigurationSum(int FeatureNum, uint8_t *FeatureEvidence, int32_t ConfigCount) {
728
0
  tprintf("F=%3d, C=", FeatureNum);
729
0
  for (int ConfigNum = 0; ConfigNum < ConfigCount; ConfigNum++) {
730
0
    tprintf("%4d", FeatureEvidence[ConfigNum]);
731
0
  }
732
0
  tprintf("\n");
733
0
}
734
735
/**
736
 * For the given feature: prune protos, compute evidence,
737
 * update Feature Evidence, Proto Evidence, and Sum of Feature
738
 * Evidence tables.
739
 * @param ClassTemplate Prototypes & tables for a class
740
 * @param FeatureNum Current feature number (for DEBUG only)
741
 * @param Feature Pointer to a feature struct
742
 * @param tables Evidence tables
743
 * @param Debug Debugger flag: 1=debugger on
744
 * @return sum of feature evidence tables
745
 */
746
int IntegerMatcher::UpdateTablesForFeature(INT_CLASS_STRUCT *ClassTemplate, BIT_VECTOR ProtoMask,
747
                                           BIT_VECTOR ConfigMask, int FeatureNum,
748
                                           const INT_FEATURE_STRUCT *Feature,
749
1.37G
                                           ScratchEvidence *tables, int Debug) {
750
1.37G
  uint32_t ConfigWord;
751
1.37G
  uint32_t ProtoWord;
752
1.37G
  uint32_t ProtoNum;
753
1.37G
  uint32_t ActualProtoNum;
754
1.37G
  uint8_t proto_byte;
755
1.37G
  int32_t proto_word_offset;
756
1.37G
  int32_t proto_offset;
757
1.37G
  PROTO_SET_STRUCT *ProtoSet;
758
1.37G
  uint32_t *ProtoPrunerPtr;
759
1.37G
  INT_PROTO_STRUCT *Proto;
760
1.37G
  int ProtoSetIndex;
761
1.37G
  uint8_t Evidence;
762
1.37G
  uint32_t XFeatureAddress;
763
1.37G
  uint32_t YFeatureAddress;
764
1.37G
  uint32_t ThetaFeatureAddress;
765
766
1.37G
  tables->ClearFeatureEvidence(ClassTemplate);
767
768
  /* Precompute Feature Address offset for Proto Pruning */
769
1.37G
  XFeatureAddress = ((Feature->X >> 2) << 1);
770
1.37G
  YFeatureAddress = (NUM_PP_BUCKETS << 1) + ((Feature->Y >> 2) << 1);
771
1.37G
  ThetaFeatureAddress = (NUM_PP_BUCKETS << 2) + ((Feature->Theta >> 2) << 1);
772
773
4.93G
  for (ProtoSetIndex = 0, ActualProtoNum = 0; ProtoSetIndex < ClassTemplate->NumProtoSets;
774
3.55G
       ProtoSetIndex++) {
775
3.55G
    ProtoSet = ClassTemplate->ProtoSets[ProtoSetIndex];
776
3.55G
    ProtoPrunerPtr = reinterpret_cast<uint32_t *>((*ProtoSet).ProtoPruner);
777
10.6G
    for (ProtoNum = 0; ProtoNum < PROTOS_PER_PROTO_SET; ProtoNum += (PROTOS_PER_PROTO_SET >> 1),
778
7.11G
        ActualProtoNum += (PROTOS_PER_PROTO_SET >> 1), ProtoMask++, ProtoPrunerPtr++) {
779
      /* Prune Protos of current Proto Set */
780
7.11G
      ProtoWord = *(ProtoPrunerPtr + XFeatureAddress);
781
7.11G
      ProtoWord &= *(ProtoPrunerPtr + YFeatureAddress);
782
7.11G
      ProtoWord &= *(ProtoPrunerPtr + ThetaFeatureAddress);
783
7.11G
      ProtoWord &= *ProtoMask;
784
785
7.11G
      if (ProtoWord != 0) {
786
3.28G
        proto_byte = ProtoWord & 0xff;
787
3.28G
        ProtoWord >>= 8;
788
3.28G
        proto_word_offset = 0;
789
9.41G
        while (ProtoWord != 0 || proto_byte != 0) {
790
11.9G
          while (proto_byte == 0) {
791
5.78G
            proto_byte = ProtoWord & 0xff;
792
5.78G
            ProtoWord >>= 8;
793
5.78G
            proto_word_offset += 8;
794
5.78G
          }
795
6.12G
          proto_offset = offset_table[proto_byte] + proto_word_offset;
796
6.12G
          proto_byte = next_table[proto_byte];
797
6.12G
          Proto = &(ProtoSet->Protos[ProtoNum + proto_offset]);
798
6.12G
          ConfigWord = Proto->Configs[0];
799
6.12G
          int32_t A3 = (((Proto->A * (Feature->X - 128)) * 2) - (Proto->B * (Feature->Y - 128)) +
800
6.12G
                        (Proto->C * 512));
801
6.12G
          int32_t M3 = ((static_cast<int8_t>(Feature->Theta - Proto->Angle)) * kIntThetaFudge) * 2;
802
803
6.12G
          if (A3 < 0) {
804
3.21G
            A3 = ~A3;
805
3.21G
          }
806
6.12G
          if (M3 < 0) {
807
2.60G
            M3 = ~M3;
808
2.60G
          }
809
6.12G
          A3 >>= mult_trunc_shift_bits_;
810
6.12G
          M3 >>= mult_trunc_shift_bits_;
811
6.12G
          if (static_cast<uint32_t>(A3) > evidence_mult_mask_) {
812
11.5M
            A3 = evidence_mult_mask_;
813
11.5M
          }
814
6.12G
          if (static_cast<uint32_t>(M3) > evidence_mult_mask_) {
815
0
            M3 = evidence_mult_mask_;
816
0
          }
817
818
6.12G
          uint32_t A4 = (A3 * A3) + (M3 * M3);
819
6.12G
          A4 >>= table_trunc_shift_bits_;
820
6.12G
          if (A4 > evidence_table_mask_) {
821
102M
            Evidence = 0;
822
6.02G
          } else {
823
6.02G
            Evidence = similarity_evidence_table_[A4];
824
6.02G
          }
825
826
6.12G
          if (PrintFeatureMatchesOn(Debug)) {
827
0
            IMDebugConfiguration(FeatureNum, ActualProtoNum + proto_offset, Evidence, ConfigWord);
828
0
          }
829
830
6.12G
          ConfigWord &= *ConfigMask;
831
832
6.12G
          uint8_t feature_evidence_index = 0;
833
6.12G
          uint8_t config_byte = 0;
834
40.1G
          while (ConfigWord != 0 || config_byte != 0) {
835
55.6G
            while (config_byte == 0) {
836
21.5G
              config_byte = ConfigWord & 0xff;
837
21.5G
              ConfigWord >>= 8;
838
21.5G
              feature_evidence_index += 8;
839
21.5G
            }
840
34.0G
            const uint8_t config_offset = offset_table[config_byte] + feature_evidence_index - 8;
841
34.0G
            config_byte = next_table[config_byte];
842
34.0G
            if (Evidence > tables->feature_evidence_[config_offset]) {
843
29.8G
              tables->feature_evidence_[config_offset] = Evidence;
844
29.8G
            }
845
34.0G
          }
846
847
6.12G
          uint8_t ProtoIndex = ClassTemplate->ProtoLengths[ActualProtoNum + proto_offset];
848
6.12G
          if (ProtoIndex > MAX_PROTO_INDEX) {
849
            // Avoid buffer overflow.
850
            // TODO: A better fix is still open.
851
1.58k
            ProtoIndex = MAX_PROTO_INDEX;
852
1.58k
          }
853
6.12G
          uint8_t *UINT8Pointer = &(tables->proto_evidence_[ActualProtoNum + proto_offset][0]);
854
26.2G
          for (; Evidence > 0 && ProtoIndex > 0; ProtoIndex--, UINT8Pointer++) {
855
20.1G
            if (Evidence > *UINT8Pointer) {
856
10.8G
              uint8_t Temp = *UINT8Pointer;
857
10.8G
              *UINT8Pointer = Evidence;
858
10.8G
              Evidence = Temp;
859
10.8G
            }
860
20.1G
          }
861
6.12G
        }
862
3.28G
      }
863
7.11G
    }
864
3.55G
  }
865
866
1.37G
  if (PrintFeatureMatchesOn(Debug)) {
867
0
    IMDebugConfigurationSum(FeatureNum, tables->feature_evidence_, ClassTemplate->NumConfigs);
868
0
  }
869
870
1.37G
  int *IntPointer = tables->sum_feature_evidence_;
871
1.37G
  uint8_t *UINT8Pointer = tables->feature_evidence_;
872
1.37G
  int SumOverConfigs = 0;
873
45.2G
  for (int ConfigNum = ClassTemplate->NumConfigs; ConfigNum > 0; ConfigNum--) {
874
43.8G
    int evidence = *UINT8Pointer++;
875
43.8G
    SumOverConfigs += evidence;
876
43.8G
    *IntPointer++ += evidence;
877
43.8G
  }
878
1.37G
  return SumOverConfigs;
879
1.37G
}
880
881
/**
882
 * Print debugging information for Configurations
883
 */
884
#ifndef GRAPHICS_DISABLED
885
void IntegerMatcher::DebugFeatureProtoError(INT_CLASS_STRUCT *ClassTemplate, BIT_VECTOR ProtoMask,
886
                                            BIT_VECTOR ConfigMask, const ScratchEvidence &tables,
887
                                            int16_t NumFeatures, int Debug) {
888
  float ProtoConfigs[MAX_NUM_CONFIGS];
889
  int ConfigNum;
890
  uint32_t ConfigWord;
891
  int ProtoSetIndex;
892
  uint16_t ProtoNum;
893
  uint8_t ProtoWordNum;
894
  PROTO_SET_STRUCT *ProtoSet;
895
  uint16_t ActualProtoNum;
896
897
  if (PrintMatchSummaryOn(Debug)) {
898
    tprintf("Configuration Mask:\n");
899
    for (ConfigNum = 0; ConfigNum < ClassTemplate->NumConfigs; ConfigNum++) {
900
      tprintf("%1d", (((*ConfigMask) >> ConfigNum) & 1));
901
    }
902
    tprintf("\n");
903
904
    tprintf("Feature Error for Configurations:\n");
905
    for (ConfigNum = 0; ConfigNum < ClassTemplate->NumConfigs; ConfigNum++) {
906
      tprintf(" %5.1f", 100.0 * (1.0 - static_cast<float>(tables.sum_feature_evidence_[ConfigNum]) /
907
                                           NumFeatures / 256.0));
908
    }
909
    tprintf("\n\n\n");
910
  }
911
912
  if (PrintMatchSummaryOn(Debug)) {
913
    tprintf("Proto Mask:\n");
914
    for (ProtoSetIndex = 0; ProtoSetIndex < ClassTemplate->NumProtoSets; ProtoSetIndex++) {
915
      ActualProtoNum = (ProtoSetIndex * PROTOS_PER_PROTO_SET);
916
      for (ProtoWordNum = 0; ProtoWordNum < 2; ProtoWordNum++, ProtoMask++) {
917
        ActualProtoNum = (ProtoSetIndex * PROTOS_PER_PROTO_SET);
918
        for (ProtoNum = 0; ((ProtoNum < (PROTOS_PER_PROTO_SET >> 1)) &&
919
                            (ActualProtoNum < ClassTemplate->NumProtos));
920
             ProtoNum++, ActualProtoNum++) {
921
          tprintf("%1d", (((*ProtoMask) >> ProtoNum) & 1));
922
        }
923
        tprintf("\n");
924
      }
925
    }
926
    tprintf("\n");
927
  }
928
929
  for (int i = 0; i < ClassTemplate->NumConfigs; i++) {
930
    ProtoConfigs[i] = 0;
931
  }
932
933
  if (PrintProtoMatchesOn(Debug)) {
934
    tprintf("Proto Evidence:\n");
935
    for (ProtoSetIndex = 0; ProtoSetIndex < ClassTemplate->NumProtoSets; ProtoSetIndex++) {
936
      ProtoSet = ClassTemplate->ProtoSets[ProtoSetIndex];
937
      ActualProtoNum = (ProtoSetIndex * PROTOS_PER_PROTO_SET);
938
      for (ProtoNum = 0;
939
           ((ProtoNum < PROTOS_PER_PROTO_SET) && (ActualProtoNum < ClassTemplate->NumProtos));
940
           ProtoNum++, ActualProtoNum++) {
941
        tprintf("P %3d =", ActualProtoNum);
942
        int temp = 0;
943
        for (uint8_t j = 0; j < ClassTemplate->ProtoLengths[ActualProtoNum]; j++) {
944
          uint8_t data = tables.proto_evidence_[ActualProtoNum][j];
945
          tprintf(" %d", data);
946
          temp += data;
947
        }
948
949
        tprintf(" = %6.4f%%\n", temp / 256.0 / ClassTemplate->ProtoLengths[ActualProtoNum]);
950
951
        ConfigWord = ProtoSet->Protos[ProtoNum].Configs[0];
952
        ConfigNum = 0;
953
        while (ConfigWord) {
954
          tprintf("%5d", ConfigWord & 1 ? temp : 0);
955
          if (ConfigWord & 1) {
956
            ProtoConfigs[ConfigNum] += temp;
957
          }
958
          ConfigNum++;
959
          ConfigWord >>= 1;
960
        }
961
        tprintf("\n");
962
      }
963
    }
964
  }
965
966
  if (PrintMatchSummaryOn(Debug)) {
967
    tprintf("Proto Error for Configurations:\n");
968
    for (ConfigNum = 0; ConfigNum < ClassTemplate->NumConfigs; ConfigNum++) {
969
      tprintf(" %5.1f", 100.0 * (1.0 - ProtoConfigs[ConfigNum] /
970
                                           ClassTemplate->ConfigLengths[ConfigNum] / 256.0));
971
    }
972
    tprintf("\n\n");
973
  }
974
975
  if (PrintProtoMatchesOn(Debug)) {
976
    tprintf("Proto Sum for Configurations:\n");
977
    for (ConfigNum = 0; ConfigNum < ClassTemplate->NumConfigs; ConfigNum++) {
978
      tprintf(" %4.1f", ProtoConfigs[ConfigNum] / 256.0);
979
    }
980
    tprintf("\n\n");
981
982
    tprintf("Proto Length for Configurations:\n");
983
    for (ConfigNum = 0; ConfigNum < ClassTemplate->NumConfigs; ConfigNum++) {
984
      tprintf(" %4.1f", static_cast<float>(ClassTemplate->ConfigLengths[ConfigNum]));
985
    }
986
    tprintf("\n\n");
987
  }
988
}
989
990
void IntegerMatcher::DisplayProtoDebugInfo(INT_CLASS_STRUCT *ClassTemplate, BIT_VECTOR ConfigMask,
991
                                           const ScratchEvidence &tables,
992
                                           bool SeparateDebugWindows) {
993
  uint16_t ProtoNum;
994
  uint16_t ActualProtoNum;
995
  PROTO_SET_STRUCT *ProtoSet;
996
  int ProtoSetIndex;
997
998
  InitIntMatchWindowIfReqd();
999
  if (SeparateDebugWindows) {
1000
    InitFeatureDisplayWindowIfReqd();
1001
    InitProtoDisplayWindowIfReqd();
1002
  }
1003
1004
  for (ProtoSetIndex = 0; ProtoSetIndex < ClassTemplate->NumProtoSets; ProtoSetIndex++) {
1005
    ProtoSet = ClassTemplate->ProtoSets[ProtoSetIndex];
1006
    ActualProtoNum = ProtoSetIndex * PROTOS_PER_PROTO_SET;
1007
    for (ProtoNum = 0;
1008
         ((ProtoNum < PROTOS_PER_PROTO_SET) && (ActualProtoNum < ClassTemplate->NumProtos));
1009
         ProtoNum++, ActualProtoNum++) {
1010
      /* Compute Average for Actual Proto */
1011
      int temp = 0;
1012
      for (uint8_t i = 0; i < ClassTemplate->ProtoLengths[ActualProtoNum]; i++) {
1013
        temp += tables.proto_evidence_[ActualProtoNum][i];
1014
      }
1015
1016
      temp /= ClassTemplate->ProtoLengths[ActualProtoNum];
1017
1018
      if ((ProtoSet->Protos[ProtoNum]).Configs[0] & (*ConfigMask)) {
1019
        DisplayIntProto(ClassTemplate, ActualProtoNum, temp / 255.0);
1020
      }
1021
    }
1022
  }
1023
}
1024
1025
void IntegerMatcher::DisplayFeatureDebugInfo(INT_CLASS_STRUCT *ClassTemplate, BIT_VECTOR ProtoMask,
1026
                                             BIT_VECTOR ConfigMask, int16_t NumFeatures,
1027
                                             const INT_FEATURE_STRUCT *Features,
1028
                                             int AdaptFeatureThreshold, int Debug,
1029
                                             bool SeparateDebugWindows) {
1030
  auto *tables = new ScratchEvidence();
1031
1032
  tables->Clear(ClassTemplate);
1033
1034
  InitIntMatchWindowIfReqd();
1035
  if (SeparateDebugWindows) {
1036
    InitFeatureDisplayWindowIfReqd();
1037
    InitProtoDisplayWindowIfReqd();
1038
  }
1039
1040
  for (int Feature = 0; Feature < NumFeatures; Feature++) {
1041
    UpdateTablesForFeature(ClassTemplate, ProtoMask, ConfigMask, Feature, &Features[Feature],
1042
                           tables, 0);
1043
1044
    /* Find Best Evidence for Current Feature */
1045
    int best = 0;
1046
    assert(ClassTemplate->NumConfigs < MAX_NUM_CONFIGS);
1047
    for (int i = 0; i < MAX_NUM_CONFIGS && i < ClassTemplate->NumConfigs; i++) {
1048
      if (tables->feature_evidence_[i] > best) {
1049
        best = tables->feature_evidence_[i];
1050
      }
1051
    }
1052
1053
    /* Update display for current feature */
1054
    if (ClipMatchEvidenceOn(Debug)) {
1055
      if (best < AdaptFeatureThreshold) {
1056
        DisplayIntFeature(&Features[Feature], 0.0);
1057
      } else {
1058
        DisplayIntFeature(&Features[Feature], 1.0);
1059
      }
1060
    } else {
1061
      DisplayIntFeature(&Features[Feature], best / 255.0);
1062
    }
1063
  }
1064
1065
  delete tables;
1066
}
1067
#endif
1068
1069
/**
1070
 * Add sum of Proto Evidences into Sum Of Feature Evidence Array
1071
 */
1072
20.6M
void ScratchEvidence::UpdateSumOfProtoEvidences(INT_CLASS_STRUCT *ClassTemplate, BIT_VECTOR ConfigMask) {
1073
20.6M
  int *IntPointer;
1074
20.6M
  uint32_t ConfigWord;
1075
20.6M
  int ProtoSetIndex;
1076
20.6M
  uint16_t ProtoNum;
1077
20.6M
  PROTO_SET_STRUCT *ProtoSet;
1078
20.6M
  int NumProtos;
1079
20.6M
  uint16_t ActualProtoNum;
1080
1081
20.6M
  NumProtos = ClassTemplate->NumProtos;
1082
1083
74.7M
  for (ProtoSetIndex = 0; ProtoSetIndex < ClassTemplate->NumProtoSets; ProtoSetIndex++) {
1084
54.0M
    ProtoSet = ClassTemplate->ProtoSets[ProtoSetIndex];
1085
54.0M
    ActualProtoNum = (ProtoSetIndex * PROTOS_PER_PROTO_SET);
1086
2.78G
    for (ProtoNum = 0; ((ProtoNum < PROTOS_PER_PROTO_SET) && (ActualProtoNum < NumProtos));
1087
2.73G
         ProtoNum++, ActualProtoNum++) {
1088
2.73G
      int temp = 0;
1089
11.2G
      for (uint8_t i = 0; i < MAX_PROTO_INDEX && i < ClassTemplate->ProtoLengths[ActualProtoNum];
1090
8.49G
           i++) {
1091
8.49G
        temp += proto_evidence_[ActualProtoNum][i];
1092
8.49G
      }
1093
1094
2.73G
      ConfigWord = ProtoSet->Protos[ProtoNum].Configs[0];
1095
2.73G
      ConfigWord &= *ConfigMask;
1096
2.73G
      IntPointer = sum_feature_evidence_;
1097
56.8G
      while (ConfigWord) {
1098
54.1G
        if (ConfigWord & 1) {
1099
12.0G
          *IntPointer += temp;
1100
12.0G
        }
1101
54.1G
        IntPointer++;
1102
54.1G
        ConfigWord >>= 1;
1103
54.1G
      }
1104
2.73G
    }
1105
54.0M
  }
1106
20.6M
}
1107
1108
/**
1109
 * Normalize Sum of Proto and Feature Evidence by dividing by the sum of
1110
 * the Feature Lengths and the Proto Lengths for each configuration.
1111
 */
1112
20.6M
void ScratchEvidence::NormalizeSums(INT_CLASS_STRUCT *ClassTemplate, int16_t NumFeatures) {
1113
  // ClassTemplate->NumConfigs can become larger than MAX_NUM_CONFIGS.
1114
684M
  for (int i = 0; i < MAX_NUM_CONFIGS && i < ClassTemplate->NumConfigs; i++) {
1115
663M
    sum_feature_evidence_[i] =
1116
663M
        (sum_feature_evidence_[i] << 8) / (NumFeatures + ClassTemplate->ConfigLengths[i]);
1117
663M
  }
1118
20.6M
}
1119
1120
/**
1121
 * Find the best match for the current class and update the Result
1122
 * with the configuration and match rating.
1123
 * @return The best normalized sum of evidences
1124
 */
1125
int IntegerMatcher::FindBestMatch(INT_CLASS_STRUCT *class_template, const ScratchEvidence &tables,
1126
20.6M
                                  UnicharRating *result) {
1127
20.6M
  int best_match = 0;
1128
20.6M
  result->config = 0;
1129
20.6M
  result->fonts.clear();
1130
20.6M
  result->fonts.reserve(class_template->NumConfigs);
1131
1132
  // Find best match.
1133
  // ClassTemplate->NumConfigs can become larger than MAX_NUM_CONFIGS.
1134
684M
  for (int c = 0; c < MAX_NUM_CONFIGS && c < class_template->NumConfigs; ++c) {
1135
663M
    int rating = tables.sum_feature_evidence_[c];
1136
663M
    if (*classify_debug_level_ > 2) {
1137
0
      tprintf("Config %d, rating=%d\n", c, rating);
1138
0
    }
1139
663M
    if (rating > best_match) {
1140
89.9M
      result->config = c;
1141
89.9M
      best_match = rating;
1142
89.9M
    }
1143
663M
    result->fonts.emplace_back(c, rating);
1144
663M
  }
1145
1146
  // Compute confidence on a Probability scale.
1147
20.6M
  result->rating = best_match / 65536.0f;
1148
1149
20.6M
  return best_match;
1150
20.6M
}
1151
1152
/**
1153
 * Applies the CN normalization factor to the given rating and returns
1154
 * the modified rating.
1155
 */
1156
float IntegerMatcher::ApplyCNCorrection(float rating, int blob_length, int normalization_factor,
1157
20.6M
                                        int matcher_multiplier) {
1158
20.6M
  int divisor = blob_length + matcher_multiplier;
1159
20.6M
  return divisor == 0
1160
20.6M
             ? 1.0f
1161
20.6M
             : (rating * blob_length + matcher_multiplier * normalization_factor / 256.0f) /
1162
20.6M
                   divisor;
1163
20.6M
}
1164
1165
} // namespace tesseract