/* 
 * Copyright 2005 - 2009 Terracotta, Inc. 
 * 
 * Licensed under the Apache License, Version 2.0 (the "License"); you may not 
 * use this file except in compliance with the License. You may obtain a copy 
 * of the License at 
 * 
 *   http://www.apache.org/licenses/LICENSE-2.0 
 *   
 * Unless required by applicable law or agreed to in writing, software 
 * distributed under the License is distributed on an "AS IS" BASIS, WITHOUT 
 * WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the 
 * License for the specific language governing permissions and limitations 
 * under the License.
 * 
 */
package com.exedosoft.plat.job.example;

import static org.quartz.DateBuilder.futureDate;
import static org.quartz.JobBuilder.newJob;
import static org.quartz.SimpleScheduleBuilder.simpleSchedule;
import static org.quartz.TriggerBuilder.newTrigger;

import org.quartz.JobDetail;
import org.quartz.Scheduler;
import org.quartz.SchedulerFactory;
import org.quartz.SimpleTrigger;
import org.quartz.DateBuilder.IntervalUnit;
import org.quartz.impl.StdSchedulerFactory;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;

/**
 * Used to test/show the clustering features of TC JobStore.
 *
 * <p>
 * All instances MUST use a different properties file, because their instance
 * Ids must be different, however all other properties should be the same.
 * </p>
 *
 * <p>
 * If you want it to clear out existing jobs & triggers, pass a command-line
 * argument called "clearJobs".
 * </p>
 *
 * <p>
 * You should probably start with a "fresh" set of tables (assuming you may
 * have some data lingering in it from other tests), since mixing data from a
 * non-clustered setup with a clustered one can be bad.
 * </p>
 *
 * <p>
 * Try killing one of the cluster instances while they are running, and see
 * that the remaining instance(s) recover the in-progress jobs. Note that
 * detection of the failure may take up to 15 or so seconds with the default
 * settings.
 * </p>
 *
 * <p>
 * Also try running it with/without the shutdown-hook plugin registered with
 * the scheduler. (org.quartz.plugins.management.ShutdownHookPlugin).
 * </p>
 *
 * <p>
 * <i>Note:</i> Never run clustering on separate machines, unless their
 * clocks are synchronized using some form of time-sync service (daemon).
 * </p>
 *
 * @author James House
 */
public class ClusterExample {

    private static Logger _log = LoggerFactory.getLogger(ClusterExample.class);


    public void run(boolean inClearJobs, boolean inScheduleJobs)
        throws Exception {

        // First we must get a reference to a scheduler
        SchedulerFactory sf = new StdSchedulerFactory();
        Scheduler sched = sf.getScheduler();

        if (inClearJobs) {
            _log.warn("***** Deleting existing jobs/triggers *****");
            sched.clear();
        }

        _log.info("------- Initialization Complete -----------");

        if (inScheduleJobs) {

            _log.info("------- Scheduling Jobs ------------------");

            String schedId = sched.getSchedulerInstanceId();

            int count = 1;

            JobDetail job = newJob(SimpleRecoveryJob.class) 
                .withIdentity("job_" + count, schedId) // put triggers in group named after the cluster node instance just to distinguish (in logging) what was scheduled from where
                .requestRecovery() // ask scheduler to re-execute this job if it was in progress when the scheduler went down...
                .build();

            
            SimpleTrigger trigger = newTrigger()
                .withIdentity("triger_" + count, schedId)
                .startAt(futureDate(1, IntervalUnit.SECOND))
                .withSchedule(simpleSchedule()
                        .withRepeatCount(20)
                        .withIntervalInSeconds(5))
                .build();
                

            _log.info(job.getKey() +
                    " will run at: " + trigger.getNextFireTime() +  
                    " and repeat: " + trigger.getRepeatCount() + 
                    " times, every " + trigger.getRepeatInterval() / 1000 + " seconds");
            sched.scheduleJob(job, trigger);

            count++;

            job = newJob(SimpleRecoveryJob.class) 
                .withIdentity("job_" + count, schedId) // put triggers in group named after the cluster node instance just to distinguish (in logging) what was scheduled from where
                .requestRecovery() // ask scheduler to re-execute this job if it was in progress when the scheduler went down...
                .build();

            trigger = newTrigger()
                .withIdentity("triger_" + count, schedId)
                .startAt(futureDate(2, IntervalUnit.SECOND))
                .withSchedule(simpleSchedule()
                        .withRepeatCount(20)
                        .withIntervalInSeconds(5))
                .build();

            _log.info(job.getKey() +
                    " will run at: " + trigger.getNextFireTime() +  
                    " and repeat: " + trigger.getRepeatCount() + 
                    " times, every " + trigger.getRepeatInterval() / 1000 + " seconds");
            sched.scheduleJob(job, trigger);

            count++;
            
            job = newJob(SimpleRecoveryStatefulJob.class) 
                .withIdentity("job_" + count, schedId) // put triggers in group named after the cluster node instance just to distinguish (in logging) what was scheduled from where
                .requestRecovery() // ask scheduler to re-execute this job if it was in progress when the scheduler went down...
                .build();
            
            trigger = newTrigger()
                .withIdentity("triger_" + count, schedId)
                .startAt(futureDate(1, IntervalUnit.SECOND))
                .withSchedule(simpleSchedule()
                        .withRepeatCount(20)
                        .withIntervalInSeconds(3))
                .build();
            
            _log.info(job.getKey() +
                    " will run at: " + trigger.getNextFireTime() +  
                    " and repeat: " + trigger.getRepeatCount() + 
                    " times, every " + trigger.getRepeatInterval() / 1000 + " seconds");
            sched.scheduleJob(job, trigger);

            count++;

            job = newJob(SimpleRecoveryJob.class) 
                .withIdentity("job_" + count, schedId) // put triggers in group named after the cluster node instance just to distinguish (in logging) what was scheduled from where
                .requestRecovery() // ask scheduler to re-execute this job if it was in progress when the scheduler went down...
                .build();

            trigger = newTrigger()
                .withIdentity("triger_" + count, schedId)
                .startAt(futureDate(1, IntervalUnit.SECOND))
                .withSchedule(simpleSchedule()
                        .withRepeatCount(20)
                        .withIntervalInSeconds(4))
                .build();
            
            _log.info(job.getKey() + " will run at: "
                    + trigger.getNextFireTime() + " & repeat: "
                    + trigger.getRepeatCount() + "/"
                    + trigger.getRepeatInterval());
            sched.scheduleJob(job, trigger);

            count++;
            
            job = newJob(SimpleRecoveryJob.class) 
                .withIdentity("job_" + count, schedId) // put triggers in group named after the cluster node instance just to distinguish (in logging) what was scheduled from where
                .requestRecovery() // ask scheduler to re-execute this job if it was in progress when the scheduler went down...
                .build();

            trigger = newTrigger()
                .withIdentity("triger_" + count, schedId)
                .startAt(futureDate(1, IntervalUnit.SECOND))
                .withSchedule(simpleSchedule()
                        .withRepeatCount(20)
                        .withIntervalInMilliseconds(4500L))
                .build();
            
            _log.info(job.getKey() + " will run at: "
                    + trigger.getNextFireTime() + " & repeat: "
                    + trigger.getRepeatCount() + "/"
                    + trigger.getRepeatInterval());
            sched.scheduleJob(job, trigger);
        }

        // jobs don't start firing until start() has been called...
        _log.info("------- Starting Scheduler ---------------");
        sched.start();
        _log.info("------- Started Scheduler ----------------");

        _log.info("------- Waiting for one hour... ----------");
        try {
            Thread.sleep(3600L * 1000L);
        } catch (Exception e) {
        }

        _log.info("------- Shutting Down --------------------");
        sched.shutdown();
        _log.info("------- Shutdown Complete ----------------");
    }

    public static void main(String[] args) throws Exception {
        boolean clearJobs = false;
        boolean scheduleJobs = true;

        for (int i = 0; i < args.length; i++) {
            if (args[i].equalsIgnoreCase("clearJobs")) {
                clearJobs = true;
            } else if (args[i].equalsIgnoreCase("dontScheduleJobs")) {
                scheduleJobs = false;
            }
        }

        ClusterExample example = new ClusterExample();
        example.run(clearJobs, scheduleJobs);
    }
}