From eca582410006443d0b101a9ae188e302f34f8a03 Mon Sep 17 00:00:00 2001 From: Jeroen Vermeulen Date: Thu, 30 Apr 2015 12:05:11 +0700 Subject: Remove trailing whitespace in C++ files. --- lm/model.hh | 40 ++++++++++++++++++++-------------------- 1 file changed, 20 insertions(+), 20 deletions(-) (limited to 'lm/model.hh') diff --git a/lm/model.hh b/lm/model.hh index c67ae2eed..b2bbe3999 100644 --- a/lm/model.hh +++ b/lm/model.hh @@ -25,7 +25,7 @@ namespace lm { namespace ngram { namespace detail { -// Should return the same results as SRI. +// Should return the same results as SRI. // ModelFacade typedefs Vocabulary so we use VocabularyT to avoid naming conflicts. template class GenericModel : public base::ModelFacade, State, VocabularyT> { private: @@ -38,7 +38,7 @@ template class GenericModel : public base::Mod /* Get the size of memory that will be mapped given ngram counts. This * does not include small non-mapped control structures, such as this class - * itself. + * itself. */ static uint64_t Size(const std::vector &counts, const Config &config = Config()); @@ -46,47 +46,47 @@ template class GenericModel : public base::Mod * files must have the format expected by this class or you'll get an * exception. So TrieModel can only load ARPA or binary created by * TrieModel. To classify binary files, call RecognizeBinary in - * lm/binary_format.hh. + * lm/binary_format.hh. */ explicit GenericModel(const char *file, const Config &config = Config()); /* Score p(new_word | in_state) and incorporate new_word into out_state. * Note that in_state and out_state must be different references: - * &in_state != &out_state. + * &in_state != &out_state. */ FullScoreReturn FullScore(const State &in_state, const WordIndex new_word, State &out_state) const; /* Slower call without in_state. Try to remember state, but sometimes it - * would cost too much memory or your decoder isn't setup properly. + * would cost too much memory or your decoder isn't setup properly. * To use this function, make an array of WordIndex containing the context * vocabulary ids in reverse order. Then, pass the bounds of the array: * [context_rbegin, context_rend). The new_word is not part of the context - * array unless you intend to repeat words. + * array unless you intend to repeat words. */ FullScoreReturn FullScoreForgotState(const WordIndex *context_rbegin, const WordIndex *context_rend, const WordIndex new_word, State &out_state) const; /* Get the state for a context. Don't use this if you can avoid it. Use * BeginSentenceState or NullContextState and extend from those. If * you're only going to use this state to call FullScore once, use - * FullScoreForgotState. + * FullScoreForgotState. * To use this function, make an array of WordIndex containing the context * vocabulary ids in reverse order. Then, pass the bounds of the array: - * [context_rbegin, context_rend). + * [context_rbegin, context_rend). */ void GetState(const WordIndex *context_rbegin, const WordIndex *context_rend, State &out_state) const; /* More efficient version of FullScore where a partial n-gram has already - * been scored. - * NOTE: THE RETURNED .rest AND .prob ARE RELATIVE TO THE .rest RETURNED BEFORE. + * been scored. + * NOTE: THE RETURNED .rest AND .prob ARE RELATIVE TO THE .rest RETURNED BEFORE. */ FullScoreReturn ExtendLeft( - // Additional context in reverse order. This will update add_rend to + // Additional context in reverse order. This will update add_rend to const WordIndex *add_rbegin, const WordIndex *add_rend, - // Backoff weights to use. + // Backoff weights to use. const float *backoff_in, // extend_left returned by a previous query. uint64_t extend_pointer, - // Length of n-gram that the pointer corresponds to. + // Length of n-gram that the pointer corresponds to. unsigned char extend_length, // Where to write additional backoffs for [extend_length + 1, min(Order() - 1, return.ngram_length)] float *backoff_out, @@ -95,17 +95,17 @@ template class GenericModel : public base::Mod /* Return probabilities minus rest costs for an array of pointers. The * first length should be the length of the n-gram to which pointers_begin - * points. + * points. */ float UnRest(const uint64_t *pointers_begin, const uint64_t *pointers_end, unsigned char first_length) const { - // Compiler should optimize this if away. + // Compiler should optimize this if away. return Search::kDifferentRest ? InternalUnRest(pointers_begin, pointers_end, first_length) : 0.0; } private: FullScoreReturn ScoreExceptBackoff(const WordIndex *const context_rbegin, const WordIndex *const context_rend, const WordIndex new_word, State &out_state) const; - // Score bigrams and above. Do not include backoff. + // Score bigrams and above. Do not include backoff. void ResumeScore(const WordIndex *context_rbegin, const WordIndex *const context_rend, unsigned char starting_order_minus_2, typename Search::Node &node, float *backoff_out, unsigned char &next_use, FullScoreReturn &ret) const; // Appears after Size in the cc file. @@ -116,7 +116,7 @@ template class GenericModel : public base::Mod float InternalUnRest(const uint64_t *pointers_begin, const uint64_t *pointers_end, unsigned char first_length) const; BinaryFormat backing_; - + VocabularyT vocab_; Search search_; @@ -124,8 +124,8 @@ template class GenericModel : public base::Mod } // namespace detail -// Instead of typedef, inherit. This allows the Model etc to be forward declared. -// Oh the joys of C and C++. +// Instead of typedef, inherit. This allows the Model etc to be forward declared. +// Oh the joys of C and C++. #define LM_COMMA() , #define LM_NAME_MODEL(name, from)\ class name : public from {\ @@ -140,7 +140,7 @@ LM_NAME_MODEL(ArrayTrieModel, detail::GenericModel LM_COMMA() SortedVocabulary>); LM_NAME_MODEL(QuantArrayTrieModel, detail::GenericModel LM_COMMA() SortedVocabulary>); -// Default implementation. No real reason for it to be the default. +// Default implementation. No real reason for it to be the default. typedef ::lm::ngram::ProbingVocabulary Vocabulary; typedef ProbingModel Model; -- cgit v1.2.3