| | import sys |
| | import os |
| | import ccxt |
| | import pandas as pd |
| | import numpy as np |
| | from datetime import datetime |
| | import ta |
| | import argparse |
| | from sklearn.ensemble import RandomForestClassifier |
| | from sklearn.model_selection import train_test_split |
| | from sklearn.metrics import accuracy_score |
| | import pickle |
| | import warnings |
| | import re |
| |
|
| | |
| | warnings.filterwarnings('ignore') |
| |
|
| | |
| | pd.set_option('display.max_columns', None) |
| | pd.set_option('display.max_rows', None) |
| | pd.set_option('display.expand_frame_repr', True) |
| |
|
| | class MLTechnicalScanner: |
| | def __init__(self, training_mode=False): |
| | self.training_mode = training_mode |
| | self.model = None |
| | self.model_file = "technical_ml_model.pkl" |
| | self.training_data_file = "training_data.csv" |
| | self.results_file = "results.txt" |
| | self.min_training_samples = 100 |
| | self.load_ml_model() |
| | |
| | |
| | self.exchanges = {} |
| | for id in ccxt.exchanges: |
| | exchange = getattr(ccxt, id) |
| | self.exchanges[id] = exchange() |
| | |
| | |
| | self.feature_columns = [ |
| | 'rsi', 'macd', 'bollinger_upper', 'bollinger_lower', |
| | 'volume_ma', 'ema_20', 'ema_50', 'adx' |
| | ] |
| | |
| | |
| | self.performance_history = pd.DataFrame(columns=[ |
| | 'timestamp', 'symbol', 'prediction', 'actual', 'profit' |
| | ]) |
| | |
| | |
| | self.training_data = pd.DataFrame(columns=self.feature_columns + ['target']) |
| | |
| | def init_results_file(self): |
| | """Initialize results file only when starting a new scan""" |
| | with open(self.results_file, 'w') as f: |
| | f.write("Scan Results Log\n") |
| | f.write("="*50 + "\n") |
| | f.write(f"Scan started at {datetime.now()}\n\n") |
| | |
| | def log_result(self, message): |
| | """Log message to results file""" |
| | try: |
| | with open(self.results_file, 'a') as f: |
| | f.write(message + '\n') |
| | except Exception as e: |
| | print(f"Error writing to results file: {str(e)}") |
| | |
| | def load_ml_model(self): |
| | """Load trained ML model if exists""" |
| | if os.path.exists(self.model_file): |
| | with open(self.model_file, 'rb') as f: |
| | self.model = pickle.load(f) |
| | msg = "Loaded trained model from file" |
| | print(msg) |
| | self.log_result(msg) |
| | else: |
| | msg = "Initializing new model" |
| | print(msg) |
| | self.log_result(msg) |
| | self.model = RandomForestClassifier(n_estimators=100, random_state=42) |
| | |
| | def save_ml_model(self): |
| | """Save trained ML model""" |
| | with open(self.model_file, 'wb') as f: |
| | pickle.dump(self.model, f) |
| | msg = "Saved model to file" |
| | print(msg) |
| | self.log_result(msg) |
| | |
| | def load_training_data(self): |
| | """Load existing training data if available""" |
| | if os.path.exists(self.training_data_file): |
| | self.training_data = pd.read_csv(self.training_data_file) |
| | msg = f"Loaded {len(self.training_data)} training samples" |
| | print(msg) |
| | self.log_result(msg) |
| | |
| | def save_training_data(self): |
| | """Save training data to file""" |
| | self.training_data.to_csv(self.training_data_file, index=False) |
| | msg = f"Saved {len(self.training_data)} training samples" |
| | print(msg) |
| | self.log_result(msg) |
| | |
| | def calculate_features(self, df): |
| | """Calculate technical indicators""" |
| | try: |
| | close = df['close'].astype(float) |
| | high = df['high'].astype(float) |
| | low = df['low'].astype(float) |
| | volume = df['volume'].astype(float) |
| | |
| | |
| | df['rsi'] = ta.momentum.rsi(close, window=14) |
| | df['macd'] = ta.trend.macd_diff(close) |
| | |
| | |
| | bollinger = ta.volatility.BollingerBands(close) |
| | df['bollinger_upper'] = bollinger.bollinger_hband() |
| | df['bollinger_lower'] = bollinger.bollinger_lband() |
| | |
| | |
| | df['volume_ma'] = volume.rolling(window=20).mean() |
| | |
| | |
| | df['ema_20'] = ta.trend.ema_indicator(close, window=20) |
| | df['ema_50'] = ta.trend.ema_indicator(close, window=50) |
| | df['adx'] = ta.trend.adx(high, low, close, window=14) |
| | |
| | return df |
| | except Exception as e: |
| | error_msg = f"Error calculating features: {str(e)}" |
| | print(error_msg) |
| | self.log_result(error_msg) |
| | return None |
| | |
| | def train_initial_model(self): |
| | """Train initial model if we have enough data""" |
| | self.load_training_data() |
| | |
| | if len(self.training_data) >= self.min_training_samples: |
| | X = self.training_data[self.feature_columns] |
| | y = self.training_data['target'] |
| | |
| | X_train, X_test, y_train, y_test = train_test_split( |
| | X, y, test_size=0.2, random_state=42 |
| | ) |
| | |
| | self.model.fit(X_train, y_train) |
| | |
| | |
| | preds = self.model.predict(X_test) |
| | accuracy = accuracy_score(y_test, preds) |
| | msg = f"Initial model trained with accuracy: {accuracy:.2f}" |
| | print(msg) |
| | self.log_result(msg) |
| | |
| | self.save_ml_model() |
| | return True |
| | else: |
| | msg = f"Not enough training data ({len(self.training_data)} samples). Need at least {self.min_training_samples}." |
| | print(msg) |
| | self.log_result(msg) |
| | return False |
| | |
| | def predict_direction(self, features): |
| | """Predict price direction using ML model""" |
| | try: |
| | if self.model is None or not hasattr(self.model, 'classes_'): |
| | return 0 |
| | |
| | features = features[self.feature_columns].values.reshape(1, -1) |
| | return self.model.predict(features)[0] |
| | except Exception as e: |
| | error_msg = f"Prediction error: {str(e)}" |
| | print(error_msg) |
| | self.log_result(error_msg) |
| | return 0 |
| | |
| | def collect_training_sample(self, symbol, exchange, timeframe='1h'): |
| | """Collect data sample for training""" |
| | try: |
| | ohlcv = exchange.fetch_ohlcv(symbol, timeframe, limit=100) |
| | if len(ohlcv) < 50: |
| | return |
| | |
| | df = pd.DataFrame(ohlcv, columns=['timestamp', 'open', 'high', 'low', 'close', 'volume']) |
| | df = self.calculate_features(df) |
| | if df is None: |
| | return |
| | |
| | current_price = df['close'].iloc[-1] |
| | future_price = df['close'].iloc[-1] |
| | |
| | price_change = future_price - current_price |
| | target = 1 if price_change > 0 else (-1 if price_change < 0 else 0) |
| | |
| | features = df.iloc[-2].copy() |
| | features['target'] = target |
| | |
| | new_row = pd.DataFrame([features]) |
| | self.training_data = pd.concat([self.training_data, new_row], ignore_index=True) |
| | msg = f"Collected training sample for {symbol}" |
| | print(msg) |
| | self.log_result(msg) |
| | |
| | if len(self.training_data) % 10 == 0: |
| | self.save_training_data() |
| | |
| | except Exception as e: |
| | error_msg = f"Error collecting training sample: {str(e)}" |
| | print(error_msg) |
| | self.log_result(error_msg) |
| | |
| | def scan_symbol(self, symbol, exchange, timeframes): |
| | """Scan symbol for trading opportunities""" |
| | try: |
| | primary_tf = timeframes[0] |
| | ohlcv = exchange.fetch_ohlcv(symbol, primary_tf, limit=100) |
| | if len(ohlcv) < 50: |
| | return |
| | |
| | df = pd.DataFrame(ohlcv, columns=['timestamp', 'open', 'high', 'low', 'close', 'volume']) |
| | df = self.calculate_features(df) |
| | if df is None: |
| | return |
| | |
| | latest = df.iloc[-1].copy() |
| | features = pd.DataFrame([latest[self.feature_columns]]) |
| | |
| | if self.training_mode: |
| | self.collect_training_sample(symbol, exchange, primary_tf) |
| | return |
| | |
| | prediction = self.predict_direction(features) |
| | |
| | |
| | ema_20 = df['ema_20'].iloc[-1] |
| | ema_50 = df['ema_50'].iloc[-1] |
| | price = df['close'].iloc[-1] |
| | |
| | uptrend = (ema_20 > ema_50) and (price > ema_20) |
| | downtrend = (ema_20 < ema_50) and (price < ema_20) |
| | |
| | if uptrend and prediction == 1: |
| | self.alert(symbol, "STRONG UPTREND", timeframes, price) |
| | elif downtrend and prediction == -1: |
| | self.alert(symbol, "STRONG DOWNTREND", timeframes, price) |
| | elif uptrend: |
| | self.alert(symbol, "UPTREND", timeframes, price) |
| | elif downtrend: |
| | self.alert(symbol, "DOWNTREND", timeframes, price) |
| | |
| | except Exception as e: |
| | error_msg = f"Error scanning {symbol}: {str(e)}" |
| | print(error_msg) |
| | self.log_result(error_msg) |
| | |
| | def alert(self, symbol, trend_type, timeframes, current_price): |
| | """Generate alert for detected trend""" |
| | timestamp = datetime.now().strftime('%Y-%m-%d %H:%M:%S') |
| | message = f"({trend_type}) detected for {symbol} at price {current_price} on {timeframes} at {timestamp}" |
| | print(message) |
| | self.log_result(message) |
| | |
| | def compare_results(self, exchange_name): |
| | """Compare previous results with current prices""" |
| | try: |
| | if not os.path.exists(self.results_file): |
| | print("No results file found to compare") |
| | return |
| | |
| | exchange = self.exchanges.get(exchange_name.lower()) |
| | if not exchange: |
| | print(f"Exchange {exchange_name} not supported") |
| | return |
| | |
| | |
| | pattern = r"\((.*?)\) detected for (.*?) at price ([\d.]+) on" |
| | |
| | with open(self.results_file, 'r') as f: |
| | lines = f.readlines() |
| | |
| | print("\n=== Price Comparison Report ===") |
| | print(f"Generated at: {datetime.now()}\n") |
| | |
| | for line in lines: |
| | match = re.search(pattern, line) |
| | if match: |
| | trend_type = match.group(1) |
| | symbol = match.group(2) |
| | old_price = float(match.group(3)) |
| | timestamp = line.split(' at ')[-1].strip() |
| | |
| | try: |
| | ticker = exchange.fetch_ticker(symbol) |
| | current_price = ticker['last'] |
| | price_change = current_price - old_price |
| | percent_change = (price_change / old_price) * 100 |
| | |
| | print(f"Symbol: {symbol}") |
| | print(f"Previous: {trend_type} at {old_price} ({timestamp})") |
| | print(f"Current: {current_price} ({datetime.now().strftime('%Y-%m-%d %H:%M:%S')})") |
| | print(f"Change: {price_change:.4f} ({percent_change:.2f}%)") |
| | print("-" * 50) |
| | except Exception as e: |
| | print(f"Error fetching current price for {symbol}: {str(e)}") |
| | continue |
| | |
| | print("\n=== End of Report ===") |
| | |
| | except Exception as e: |
| | print(f"Error comparing results: {str(e)}") |
| |
|
| | |
| | if __name__ == "__main__": |
| | parser = argparse.ArgumentParser() |
| | parser.add_argument("-e", "--exchange", help="Exchange name", required=False) |
| | parser.add_argument("-f", "--filter", help="Asset filter", required=False) |
| | parser.add_argument("-tf", "--timeframes", help="Timeframes to scan (comma separated)", required=False) |
| | parser.add_argument("--train", help="Run in training mode", action="store_true") |
| | parser.add_argument("--compare", help="Compare previous results with current prices", action="store_true") |
| | args = parser.parse_args() |
| | |
| | if args.compare: |
| | scanner = MLTechnicalScanner() |
| | if args.exchange: |
| | scanner.compare_results(args.exchange) |
| | else: |
| | print("Please specify an exchange with -e/--exchange when using --compare") |
| | sys.exit(0) |
| | |
| | if not all([args.exchange, args.filter, args.timeframes]): |
| | print("Error: --exchange, --filter, and --timeframes are required when not using --compare") |
| | sys.exit(1) |
| | |
| | scanner = MLTechnicalScanner(training_mode=args.train) |
| | |
| | |
| | scanner.init_results_file() |
| | |
| | exchange = scanner.exchanges.get(args.exchange.lower()) |
| | if not exchange: |
| | error_msg = f"Exchange {args.exchange} not supported" |
| | print(error_msg) |
| | scanner.log_result(error_msg) |
| | sys.exit(1) |
| | |
| | try: |
| | markets = exchange.fetch_markets() |
| | except Exception as e: |
| | error_msg = f"Error fetching markets: {str(e)}" |
| | print(error_msg) |
| | scanner.log_result(error_msg) |
| | sys.exit(1) |
| | |
| | symbols = [ |
| | m['id'] for m in markets |
| | if m['active'] and args.filter in m['id'] |
| | ] |
| | |
| | if not symbols: |
| | error_msg = f"No symbols found matching filter {args.filter}" |
| | print(error_msg) |
| | scanner.log_result(error_msg) |
| | sys.exit(1) |
| | |
| | if args.train: |
| | train_msg = f"Running in training mode for {len(symbols)} symbols" |
| | print(train_msg) |
| | scanner.log_result(train_msg) |
| | for symbol in symbols: |
| | scanner.collect_training_sample(symbol, exchange) |
| | |
| | if scanner.train_initial_model(): |
| | success_msg = "Training completed successfully" |
| | print(success_msg) |
| | scanner.log_result(success_msg) |
| | else: |
| | fail_msg = "Not enough data collected for training" |
| | print(fail_msg) |
| | scanner.log_result(fail_msg) |
| | sys.exit(0) |
| | |
| | if not hasattr(scanner.model, 'classes_'): |
| | warn_msg = "Warning: No trained model available. Running with basic scanning only." |
| | print(warn_msg) |
| | scanner.log_result(warn_msg) |
| | |
| | timeframes = args.timeframes.split(',') |
| | scan_msg = f"Scanning {len(symbols)} symbols on timeframes {timeframes}" |
| | print(scan_msg) |
| | scanner.log_result(scan_msg) |
| | |
| | for symbol in symbols: |
| | scanner.scan_symbol(symbol, exchange, timeframes) |
| | |
| | |
| | end_msg = f"\nScan completed at {datetime.now()}" |
| | print(end_msg) |
| | scanner.log_result(end_msg) |