| #include "OxLM.h" |
|
|
| #include <boost/archive/binary_iarchive.hpp> |
| #include <boost/archive/binary_oarchive.hpp> |
| #include <boost/filesystem.hpp> |
| #include <boost/functional/hash.hpp> |
|
|
| #include "moses/FactorCollection.h" |
| #include "moses/InputType.h" |
| #include "moses/TranslationTask.h" |
|
|
| using namespace std; |
| using namespace oxlm; |
|
|
| namespace Moses |
| { |
|
|
| template<class Model> |
| OxLM<Model>::OxLM(const string &line) |
| : LanguageModelSingleFactor(line), normalized(true), |
| posBackOff(false), posFactorType(1), |
| persistentCache(false) |
| { |
| ReadParameters(); |
|
|
| FactorCollection &factorCollection = FactorCollection::Instance(); |
|
|
| |
| m_sentenceStart = factorCollection.AddFactor(Output, m_factorType, BOS_); |
| m_sentenceStartWord[m_factorType] = m_sentenceStart; |
|
|
| m_sentenceEnd = factorCollection.AddFactor(Output, m_factorType, EOS_); |
| m_sentenceEndWord[m_factorType] = m_sentenceEnd; |
|
|
| cacheHits = totalHits = 0; |
| } |
|
|
|
|
| template<class Model> |
| OxLM<Model>::~OxLM() |
| { |
| if (persistentCache) { |
| if (cache.get()) { |
| string cache_file = m_filePath + ".phrases.cache.bin"; |
| savePersistentCache(cache_file); |
| } |
|
|
| double cache_hit_ratio = 100.0 * cacheHits / totalHits; |
| cerr << "Cache hit ratio: " << cache_hit_ratio << endl; |
| } |
| } |
|
|
|
|
| template<class Model> |
| void OxLM<Model>::SetParameter(const string& key, const string& value) |
| { |
| if (key == "normalized") { |
| normalized = Scan<bool>(value); |
| } else if (key == "persistent-cache") { |
| persistentCache = Scan<bool>(value); |
| } else if (key == "normalized") { |
| normalized = Scan<bool>(value); |
| } else if (key == "pos-back-off") { |
| posBackOff = Scan<bool>(value); |
| } else if (key == "pos-factor-type") { |
| posFactorType = Scan<FactorType>(value); |
| } else { |
| LanguageModelSingleFactor::SetParameter(key, value); |
| } |
| } |
|
|
| template<class Model> |
| void OxLM<Model>::Load(AllOptions::ptr const& opts) |
| { |
| model.load(m_filePath); |
|
|
| boost::shared_ptr<Vocabulary> vocab = model.getVocab(); |
| mapper = boost::make_shared<OxLMMapper>(vocab, posBackOff, posFactorType); |
|
|
| kSTART = vocab->convert("<s>"); |
| kSTOP = vocab->convert("</s>"); |
| kUNKNOWN = vocab->convert("<unk>"); |
|
|
| size_t ngram_order = model.getConfig()->ngram_order; |
| UTIL_THROW_IF2( |
| m_nGramOrder != ngram_order, |
| "Wrong order for OxLM: LM has " << ngram_order << ", but Moses expects " << m_nGramOrder); |
| } |
|
|
| template<class Model> |
| double OxLM<Model>::GetScore(int word, const vector<int>& context) const |
| { |
| if (normalized) { |
| return model.getLogProb(word, context); |
| } else { |
| return model.getUnnormalizedScore(word, context); |
| } |
| } |
|
|
| template<class Model> |
| LMResult OxLM<Model>::GetValue( |
| const vector<const Word*> &contextFactor, State* finalState) const |
| { |
| if (!cache.get()) { |
| cache.reset(new QueryCache()); |
| string cache_file = m_filePath + ".phrases.cache.bin"; |
| loadPersistentCache(cache_file); |
| } |
|
|
| vector<int> context; |
| int word; |
| mapper->convert(contextFactor, context, word); |
|
|
| size_t context_width = m_nGramOrder - 1; |
| if (!context.empty() && context.back() == kSTART) { |
| context.resize(context_width, kSTART); |
| } else { |
| context.resize(context_width, kUNKNOWN); |
| } |
|
|
| double score; |
| if (persistentCache) { |
| ++totalHits; |
| NGram query(word, context); |
| pair<double, bool> ret = cache->get(query); |
| if (ret.second) { |
| score = ret.first; |
| ++cacheHits; |
| } else { |
| score = GetScore(word, context); |
| cache->put(query, score); |
| } |
| } else { |
| score = GetScore(word, context); |
| } |
|
|
| LMResult ret; |
| ret.score = score; |
| ret.unknown = (word == kUNKNOWN); |
|
|
| |
| size_t seed = 0; |
| boost::hash_combine(seed, word); |
| for (size_t i = 0; i < context.size() && i < context_width - 1; ++i) { |
| int id = context[i]; |
| boost::hash_combine(seed, id); |
| } |
|
|
| (*finalState) = (State*) seed; |
| return ret; |
| } |
|
|
| template<class Model> |
| void OxLM<Model>::loadPersistentCache(const string& cache_file) const |
| { |
| if (boost::filesystem::exists(cache_file)) { |
| ifstream f(cache_file); |
| boost::archive::binary_iarchive iar(f); |
| cerr << "Loading n-gram probability cache from " << cache_file << endl; |
| iar >> *cache; |
| cerr << "Done loading " << cache->size() |
| << " n-gram probabilities..." << endl; |
| } else { |
| cerr << "Cache file not found" << endl; |
| } |
| } |
|
|
| template<class Model> |
| void OxLM<Model>::savePersistentCache(const string& cache_file) const |
| { |
| ofstream f(cache_file); |
| boost::archive::binary_oarchive oar(f); |
| cerr << "Saving persistent cache to " << cache_file << endl; |
| oar << *cache; |
| cerr << "Done saving " << cache->size() |
| << " n-gram probabilities..." << endl; |
| } |
|
|
| template<class Model> |
| void OxLM<Model>::InitializeForInput(ttasksptr const& ttask) |
| { |
| const InputType& source = *ttask->GetSource(); |
| LanguageModelSingleFactor::InitializeForInput(ttask); |
|
|
| if (persistentCache) { |
| if (!cache.get()) { |
| cache.reset(new QueryCache()); |
| } |
|
|
| int sentence_id = source.GetTranslationId(); |
| string cache_file = m_filePath + "." + to_string(sentence_id) + ".cache.bin"; |
| loadPersistentCache(cache_file); |
| } |
| } |
|
|
| template<class Model> |
| void OxLM<Model>::CleanUpAfterSentenceProcessing(const InputType& source) |
| { |
| |
| model.clearCache(); |
|
|
| if (persistentCache) { |
| int sentence_id = source.GetTranslationId(); |
| string cache_file = m_filePath + "." + to_string(sentence_id) + ".cache.bin"; |
| savePersistentCache(cache_file); |
|
|
| cache->clear(); |
| } |
|
|
| LanguageModelSingleFactor::CleanUpAfterSentenceProcessing(source); |
| } |
|
|
| template class OxLM<LM>; |
| template class OxLM<FactoredLM>; |
| template class OxLM<FactoredMaxentLM>; |
| template class OxLM<FactoredTreeLM>; |
|
|
| } |
|
|
|
|
|
|
|
|