-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathstormscale.py
More file actions
141 lines (123 loc) · 4.87 KB
/
stormscale.py
File metadata and controls
141 lines (123 loc) · 4.87 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
import os
import time
import json
import locust
import logging
import pandas as pd
import requests
import numpy as np
from locust import HttpUser, task, between
from sklearn.ensemble import RandomForestRegressor
from sklearn.model_selection import train_test_split
import psutil
import matplotlib.pyplot as plt
import threading
import tensorflow as tf
from keras.models import Sequential
from keras.layers import Dense, LSTM, Dropout
from flask import Flask, jsonify, request
import socket
import seaborn as sns
import plotly.express as px
# Set up logging
logging.basicConfig(level=logging.INFO, format='%(asctime)s - %(levelname)s - %(message)s')
class StormScaleUser(HttpUser):
wait_time = between(1, 5)
def on_start(self):
self.login()
def login(self):
payload = {"username": "testuser", "password": "password123"}
response = self.client.post("/api/login", json=payload)
if response.status_code == 200:
self.token = response.json().get("token")
logging.info("Login successful")
else:
logging.error("Login failed")
self.token = None
@task(1)
def create_dataverse(self):
if self.token:
headers = {"Authorization": f"Bearer {self.token}"}
data = {"name": "TestDataverse", "description": "Performance Testing Dataverse"}
response = self.client.post("/api/dataverse", headers=headers, json=data)
if response.status_code == 201:
logging.info("Dataverse created successfully")
else:
logging.error("Failed to create Dataverse")
@task(2)
def create_dataset(self):
if self.token:
headers = {"Authorization": f"Bearer {self.token}"}
data = {"title": "TestDataset", "description": "AI-powered Performance Test Dataset"}
response = self.client.post("/api/dataset", headers=headers, json=data)
if response.status_code == 201:
logging.info("Dataset created successfully")
else:
logging.error("Failed to create Dataset")
# AI-Based Performance Prediction
class AIModel:
def __init__(self):
self.model = Sequential([
LSTM(100, activation='relu', return_sequences=True, input_shape=(10, 4)),
Dropout(0.2),
LSTM(50, activation='relu'),
Dense(1)
])
self.model.compile(optimizer='adam', loss='mse')
def train(self, X, y):
self.model.fit(X, y, epochs=10, batch_size=32, verbose=1)
def predict(self, X_new):
return self.model.predict(np.array([X_new]))
# AI-driven Anomaly Detection
def detect_anomalies(data):
threshold = data['response_time'].mean() + (2 * data['response_time'].std())
anomalies = data[data['response_time'] > threshold]
return anomalies
# System Resource Monitoring
def monitor_system_resources():
while True:
cpu = psutil.cpu_percent(interval=1)
memory = psutil.virtual_memory().percent
disk = psutil.disk_usage('/').percent
network = psutil.net_io_counters().bytes_sent + psutil.net_io_counters().bytes_recv
logging.info(f"CPU: {cpu}% | Memory: {memory}% | Disk: {disk}% | Network: {network} bytes")
time.sleep(5)
# API for Real-Time Performance Monitoring
app = Flask(__name__)
@app.route('/status', methods=['GET'])
def get_system_status():
cpu = psutil.cpu_percent(interval=1)
memory = psutil.virtual_memory().percent
disk = psutil.disk_usage('/').percent
return jsonify({"cpu": cpu, "memory": memory, "disk": disk})
@app.route('/predict', methods=['POST'])
def predict_performance():
data = request.get_json()
X_new = np.array(data['features'])
model = AIModel()
prediction = model.predict(X_new)
return jsonify({"predicted_performance": prediction.tolist()})
@app.route('/detect-anomalies', methods=['GET'])
def anomaly_detection():
df = pd.read_csv("performance_data.csv")
anomalies = detect_anomalies(df)
return anomalies.to_json()
# Auto-scaling performance testing
def auto_scale_tests():
while True:
current_cpu = psutil.cpu_percent()
if current_cpu < 50:
os.system("locust --users=100 --spawn-rate=10 --run-time=2m")
elif current_cpu < 80:
os.system("locust --users=50 --spawn-rate=5 --run-time=2m")
time.sleep(60)
# Network Monitoring
def get_local_ip():
return "127.0.0.1"
if __name__ == "__main__":
logging.info("Starting StormScale - AI-Powered Performance Testing Tool")
threading.Thread(target=monitor_system_resources, daemon=True).start()
threading.Thread(target=lambda: app.run(host='0.0.0.0', port=5000), daemon=True).start()
threading.Thread(target=auto_scale_tests, daemon=True).start()
logging.info("System monitoring API available at http://127.0.0.1:5000/status")
os.system("locust -f stormscale.py")