/*
 * This file Copyright (c) 2015. Walle.
 * (http://www.wallellen.com). All rights reserved.
 *
 *
 * This file is dual-licensed under both the
 * Walle Agreement (WA) and the GNU General Public License.
 * You may elect to use one or the other of these licenses.
 *
 * This file is distributed in the hope that it will be
 * useful, but AS-IS and WITHOUT ANY WARRANTY; without even the
 * implied warranty of MERCHANTABILITY or FITNESS FOR A
 * PARTICULAR PURPOSE, TITLE, or NONINFRINGEMENT.
 * Redistribution, except as permitted by whichever of the GPL
 * or WA you select, is prohibited.
 *
 * 1. For the GPL license (GPL), you can redistribute and/or
 * modify this file under the terms of the GNU General
 * Public License, Version 3, as published by the Free Software
 * Foundation.  You should have received a copy of the GNU
 * General Public License, Version 3 along with this program;
 * if not, write to the Free Software Foundation, Inc., 51
 * Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
 *
 * 2. For the Walle Agreement (WA), this file
 * and the accompanying materials are made available under the
 * terms of the WA which accompanies this distribution, and
 * is available at http://www.wallellen.com/agreement.html
 *
 *
 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER
 */

/**
 *
 */
package org.wltea.analyzer.solr;

import org.apache.lucene.analysis.Tokenizer;
import org.apache.solr.analysis.BaseTokenizerFactory;
import org.wltea.analyzer.lucene.IKTokenizer;

import java.io.Reader;
import java.util.Map;

/**
 * 实现Solr1.4分词器接口
 * 基于IKTokenizer的实现
 *
 * @author 林良益、李良杰
 */
public final class IKTokenizerFactory extends BaseTokenizerFactory {

    private boolean isMaxWordLength = false;

    /**
     * IK分词器Solr TokenizerFactory接口实现类
     * 默认最细粒度切分算法
     */
    public IKTokenizerFactory() {
    }

    /*
     * (non-Javadoc)
     * @see org.apache.solr.analysis.BaseTokenizerFactory#init(java.util.Map)
     */
    public void init(Map<String, String> args) {
        String _arg = args.get("isMaxWordLength");
        isMaxWordLength = Boolean.parseBoolean(_arg);
    }

    /*
     * (non-Javadoc)
     * @see org.apache.solr.analysis.TokenizerFactory#create(java.io.Reader)
     */
    public Tokenizer create(Reader reader) {
        return new IKTokenizer(reader, isMaxWordLength());
    }

    public void setMaxWordLength(boolean isMaxWordLength) {
        this.isMaxWordLength = isMaxWordLength;
    }

    public boolean isMaxWordLength() {
        return isMaxWordLength;
    }

}
