Optimize metrics import and split off metrics import functions (untested)
This commit is contained in:
@@ -30,10 +30,31 @@ CREATE TABLE IF NOT EXISTS products (
|
|||||||
INDEX idx_brand (brand)
|
INDEX idx_brand (brand)
|
||||||
);
|
);
|
||||||
|
|
||||||
|
-- Temporary tables for batch metrics processing
|
||||||
|
CREATE TABLE IF NOT EXISTS temp_sales_metrics (
|
||||||
|
product_id BIGINT NOT NULL,
|
||||||
|
daily_sales_avg DECIMAL(10,3),
|
||||||
|
weekly_sales_avg DECIMAL(10,3),
|
||||||
|
monthly_sales_avg DECIMAL(10,3),
|
||||||
|
total_revenue DECIMAL(10,3),
|
||||||
|
avg_margin_percent DECIMAL(10,3),
|
||||||
|
first_sale_date DATE,
|
||||||
|
last_sale_date DATE,
|
||||||
|
PRIMARY KEY (product_id)
|
||||||
|
);
|
||||||
|
|
||||||
|
CREATE TABLE IF NOT EXISTS temp_purchase_metrics (
|
||||||
|
product_id BIGINT NOT NULL,
|
||||||
|
avg_lead_time_days INT,
|
||||||
|
last_purchase_date DATE,
|
||||||
|
last_received_date DATE,
|
||||||
|
PRIMARY KEY (product_id)
|
||||||
|
);
|
||||||
|
|
||||||
-- New table for product metrics
|
-- New table for product metrics
|
||||||
CREATE TABLE IF NOT EXISTS product_metrics (
|
CREATE TABLE IF NOT EXISTS product_metrics (
|
||||||
product_id BIGINT NOT NULL,
|
product_id BIGINT NOT NULL,
|
||||||
last_calculated_at TIMESTAMP NOT NULL,
|
last_calculated_at TIMESTAMP NOT NULL DEFAULT CURRENT_TIMESTAMP,
|
||||||
-- Sales velocity metrics
|
-- Sales velocity metrics
|
||||||
daily_sales_avg DECIMAL(10,3),
|
daily_sales_avg DECIMAL(10,3),
|
||||||
weekly_sales_avg DECIMAL(10,3),
|
weekly_sales_avg DECIMAL(10,3),
|
||||||
@@ -57,6 +78,10 @@ CREATE TABLE IF NOT EXISTS product_metrics (
|
|||||||
FOREIGN KEY (product_id) REFERENCES products(product_id) ON DELETE CASCADE
|
FOREIGN KEY (product_id) REFERENCES products(product_id) ON DELETE CASCADE
|
||||||
);
|
);
|
||||||
|
|
||||||
|
-- Optimized indexes for metrics calculations
|
||||||
|
CREATE INDEX idx_orders_metrics ON orders (product_id, date, canceled, quantity, price);
|
||||||
|
CREATE INDEX idx_purchase_orders_metrics ON purchase_orders (product_id, date, status, ordered, received);
|
||||||
|
|
||||||
-- New table for time-based aggregates
|
-- New table for time-based aggregates
|
||||||
CREATE TABLE IF NOT EXISTS product_time_aggregates (
|
CREATE TABLE IF NOT EXISTS product_time_aggregates (
|
||||||
product_id BIGINT NOT NULL,
|
product_id BIGINT NOT NULL,
|
||||||
|
|||||||
359
inventory-server/scripts/calculate-metrics.js
Normal file
359
inventory-server/scripts/calculate-metrics.js
Normal file
@@ -0,0 +1,359 @@
|
|||||||
|
const mysql = require('mysql2/promise');
|
||||||
|
const path = require('path');
|
||||||
|
require('dotenv').config({ path: path.resolve(__dirname, '..', '.env') });
|
||||||
|
|
||||||
|
// Helper function to output progress
|
||||||
|
function outputProgress(data) {
|
||||||
|
console.log(JSON.stringify(data));
|
||||||
|
}
|
||||||
|
|
||||||
|
// Helper function to log errors
|
||||||
|
function logError(error, context) {
|
||||||
|
console.error(JSON.stringify({
|
||||||
|
status: 'error',
|
||||||
|
error: error.message || error,
|
||||||
|
context
|
||||||
|
}));
|
||||||
|
}
|
||||||
|
|
||||||
|
// Database configuration
|
||||||
|
const dbConfig = {
|
||||||
|
host: process.env.DB_HOST,
|
||||||
|
user: process.env.DB_USER,
|
||||||
|
password: process.env.DB_PASSWORD,
|
||||||
|
database: process.env.DB_NAME,
|
||||||
|
waitForConnections: true,
|
||||||
|
connectionLimit: 10,
|
||||||
|
queueLimit: 0
|
||||||
|
};
|
||||||
|
|
||||||
|
async function calculateMetrics() {
|
||||||
|
let pool;
|
||||||
|
try {
|
||||||
|
pool = mysql.createPool(dbConfig);
|
||||||
|
const connection = await pool.getConnection();
|
||||||
|
|
||||||
|
try {
|
||||||
|
// Create temporary tables for metrics calculations
|
||||||
|
outputProgress({
|
||||||
|
status: 'running',
|
||||||
|
operation: 'Creating temporary tables',
|
||||||
|
percentage: '0'
|
||||||
|
});
|
||||||
|
|
||||||
|
await connection.query(`
|
||||||
|
CREATE TABLE IF NOT EXISTS temp_sales_metrics (
|
||||||
|
product_id INT PRIMARY KEY,
|
||||||
|
total_quantity_sold INT DEFAULT 0,
|
||||||
|
total_revenue DECIMAL(10,2) DEFAULT 0.00,
|
||||||
|
average_price DECIMAL(10,2) DEFAULT 0.00,
|
||||||
|
last_sale_date DATE,
|
||||||
|
sales_rank INT
|
||||||
|
);
|
||||||
|
|
||||||
|
CREATE TABLE IF NOT EXISTS temp_purchase_metrics (
|
||||||
|
product_id INT PRIMARY KEY,
|
||||||
|
total_quantity_purchased INT DEFAULT 0,
|
||||||
|
total_cost DECIMAL(10,2) DEFAULT 0.00,
|
||||||
|
average_cost DECIMAL(10,2) DEFAULT 0.00,
|
||||||
|
last_purchase_date DATE,
|
||||||
|
purchase_rank INT
|
||||||
|
);
|
||||||
|
|
||||||
|
TRUNCATE TABLE temp_sales_metrics;
|
||||||
|
TRUNCATE TABLE temp_purchase_metrics;
|
||||||
|
`);
|
||||||
|
|
||||||
|
// Calculate sales metrics
|
||||||
|
outputProgress({
|
||||||
|
status: 'running',
|
||||||
|
operation: 'Calculating sales metrics',
|
||||||
|
percentage: '20'
|
||||||
|
});
|
||||||
|
|
||||||
|
await connection.query(`
|
||||||
|
INSERT INTO temp_sales_metrics (
|
||||||
|
product_id,
|
||||||
|
total_quantity_sold,
|
||||||
|
total_revenue,
|
||||||
|
average_price,
|
||||||
|
last_sale_date
|
||||||
|
)
|
||||||
|
SELECT
|
||||||
|
product_id,
|
||||||
|
SUM(quantity) as total_quantity_sold,
|
||||||
|
SUM((price - COALESCE(discount, 0)) * quantity) as total_revenue,
|
||||||
|
AVG(price - COALESCE(discount, 0)) as average_price,
|
||||||
|
MAX(date) as last_sale_date
|
||||||
|
FROM orders
|
||||||
|
WHERE canceled = 0
|
||||||
|
GROUP BY product_id;
|
||||||
|
|
||||||
|
UPDATE temp_sales_metrics
|
||||||
|
SET sales_rank = (
|
||||||
|
SELECT rank
|
||||||
|
FROM (
|
||||||
|
SELECT
|
||||||
|
product_id,
|
||||||
|
RANK() OVER (ORDER BY total_revenue DESC) as rank
|
||||||
|
FROM temp_sales_metrics
|
||||||
|
) rankings
|
||||||
|
WHERE rankings.product_id = temp_sales_metrics.product_id
|
||||||
|
);
|
||||||
|
`);
|
||||||
|
|
||||||
|
// Calculate purchase metrics
|
||||||
|
outputProgress({
|
||||||
|
status: 'running',
|
||||||
|
operation: 'Calculating purchase metrics',
|
||||||
|
percentage: '40'
|
||||||
|
});
|
||||||
|
|
||||||
|
await connection.query(`
|
||||||
|
INSERT INTO temp_purchase_metrics (
|
||||||
|
product_id,
|
||||||
|
total_quantity_purchased,
|
||||||
|
total_cost,
|
||||||
|
average_cost,
|
||||||
|
last_purchase_date
|
||||||
|
)
|
||||||
|
SELECT
|
||||||
|
product_id,
|
||||||
|
SUM(received) as total_quantity_purchased,
|
||||||
|
SUM(cost_price * received) as total_cost,
|
||||||
|
AVG(cost_price) as average_cost,
|
||||||
|
MAX(received_date) as last_purchase_date
|
||||||
|
FROM purchase_orders
|
||||||
|
WHERE status = 'closed' AND received > 0
|
||||||
|
GROUP BY product_id;
|
||||||
|
|
||||||
|
UPDATE temp_purchase_metrics
|
||||||
|
SET purchase_rank = (
|
||||||
|
SELECT rank
|
||||||
|
FROM (
|
||||||
|
SELECT
|
||||||
|
product_id,
|
||||||
|
RANK() OVER (ORDER BY total_cost DESC) as rank
|
||||||
|
FROM temp_purchase_metrics
|
||||||
|
) rankings
|
||||||
|
WHERE rankings.product_id = temp_purchase_metrics.product_id
|
||||||
|
);
|
||||||
|
`);
|
||||||
|
|
||||||
|
// Update product metrics
|
||||||
|
outputProgress({
|
||||||
|
status: 'running',
|
||||||
|
operation: 'Updating product metrics',
|
||||||
|
percentage: '60'
|
||||||
|
});
|
||||||
|
|
||||||
|
await connection.query(`
|
||||||
|
INSERT INTO product_metrics (
|
||||||
|
product_id,
|
||||||
|
total_quantity_sold,
|
||||||
|
total_revenue,
|
||||||
|
average_price,
|
||||||
|
total_quantity_purchased,
|
||||||
|
total_cost,
|
||||||
|
average_cost,
|
||||||
|
profit_margin,
|
||||||
|
turnover_rate,
|
||||||
|
last_sale_date,
|
||||||
|
last_purchase_date,
|
||||||
|
sales_rank,
|
||||||
|
purchase_rank,
|
||||||
|
last_calculated_at
|
||||||
|
)
|
||||||
|
SELECT
|
||||||
|
p.product_id,
|
||||||
|
COALESCE(s.total_quantity_sold, 0),
|
||||||
|
COALESCE(s.total_revenue, 0.00),
|
||||||
|
COALESCE(s.average_price, 0.00),
|
||||||
|
COALESCE(po.total_quantity_purchased, 0),
|
||||||
|
COALESCE(po.total_cost, 0.00),
|
||||||
|
COALESCE(po.average_cost, 0.00),
|
||||||
|
CASE
|
||||||
|
WHEN COALESCE(s.total_revenue, 0) = 0 THEN 0
|
||||||
|
ELSE ((s.total_revenue - po.total_cost) / s.total_revenue) * 100
|
||||||
|
END as profit_margin,
|
||||||
|
CASE
|
||||||
|
WHEN COALESCE(po.total_quantity_purchased, 0) = 0 THEN 0
|
||||||
|
ELSE (s.total_quantity_sold / po.total_quantity_purchased) * 100
|
||||||
|
END as turnover_rate,
|
||||||
|
s.last_sale_date,
|
||||||
|
po.last_purchase_date,
|
||||||
|
s.sales_rank,
|
||||||
|
po.purchase_rank,
|
||||||
|
NOW()
|
||||||
|
FROM products p
|
||||||
|
LEFT JOIN temp_sales_metrics s ON p.product_id = s.product_id
|
||||||
|
LEFT JOIN temp_purchase_metrics po ON p.product_id = po.product_id
|
||||||
|
ON DUPLICATE KEY UPDATE
|
||||||
|
total_quantity_sold = VALUES(total_quantity_sold),
|
||||||
|
total_revenue = VALUES(total_revenue),
|
||||||
|
average_price = VALUES(average_price),
|
||||||
|
total_quantity_purchased = VALUES(total_quantity_purchased),
|
||||||
|
total_cost = VALUES(total_cost),
|
||||||
|
average_cost = VALUES(average_cost),
|
||||||
|
profit_margin = VALUES(profit_margin),
|
||||||
|
turnover_rate = VALUES(turnover_rate),
|
||||||
|
last_sale_date = VALUES(last_sale_date),
|
||||||
|
last_purchase_date = VALUES(last_purchase_date),
|
||||||
|
sales_rank = VALUES(sales_rank),
|
||||||
|
purchase_rank = VALUES(purchase_rank),
|
||||||
|
last_calculated_at = VALUES(last_calculated_at);
|
||||||
|
`);
|
||||||
|
|
||||||
|
// Calculate ABC classification
|
||||||
|
outputProgress({
|
||||||
|
status: 'running',
|
||||||
|
operation: 'Calculating ABC classification',
|
||||||
|
percentage: '80'
|
||||||
|
});
|
||||||
|
|
||||||
|
await connection.query(`
|
||||||
|
WITH revenue_percentiles AS (
|
||||||
|
SELECT
|
||||||
|
product_id,
|
||||||
|
total_revenue,
|
||||||
|
PERCENT_RANK() OVER (ORDER BY total_revenue DESC) as revenue_percentile
|
||||||
|
FROM product_metrics
|
||||||
|
WHERE total_revenue > 0
|
||||||
|
)
|
||||||
|
UPDATE product_metrics pm
|
||||||
|
JOIN revenue_percentiles rp ON pm.product_id = rp.product_id
|
||||||
|
SET pm.abc_class =
|
||||||
|
CASE
|
||||||
|
WHEN rp.revenue_percentile < 0.2 THEN 'A'
|
||||||
|
WHEN rp.revenue_percentile < 0.5 THEN 'B'
|
||||||
|
ELSE 'C'
|
||||||
|
END;
|
||||||
|
`);
|
||||||
|
|
||||||
|
// Calculate time-based aggregates
|
||||||
|
outputProgress({
|
||||||
|
status: 'running',
|
||||||
|
operation: 'Calculating time aggregates',
|
||||||
|
percentage: '90'
|
||||||
|
});
|
||||||
|
|
||||||
|
await connection.query(`
|
||||||
|
TRUNCATE TABLE product_time_aggregates;
|
||||||
|
|
||||||
|
-- Daily aggregates
|
||||||
|
INSERT INTO product_time_aggregates (product_id, period_type, period_start, quantity_sold, revenue)
|
||||||
|
SELECT
|
||||||
|
product_id,
|
||||||
|
'daily' as period_type,
|
||||||
|
DATE(date) as period_start,
|
||||||
|
SUM(quantity) as quantity_sold,
|
||||||
|
SUM((price - COALESCE(discount, 0)) * quantity) as revenue
|
||||||
|
FROM orders
|
||||||
|
WHERE canceled = 0
|
||||||
|
GROUP BY product_id, DATE(date);
|
||||||
|
|
||||||
|
-- Weekly aggregates
|
||||||
|
INSERT INTO product_time_aggregates (product_id, period_type, period_start, quantity_sold, revenue)
|
||||||
|
SELECT
|
||||||
|
product_id,
|
||||||
|
'weekly' as period_type,
|
||||||
|
DATE(DATE_SUB(date, INTERVAL WEEKDAY(date) DAY)) as period_start,
|
||||||
|
SUM(quantity) as quantity_sold,
|
||||||
|
SUM((price - COALESCE(discount, 0)) * quantity) as revenue
|
||||||
|
FROM orders
|
||||||
|
WHERE canceled = 0
|
||||||
|
GROUP BY product_id, DATE(DATE_SUB(date, INTERVAL WEEKDAY(date) DAY));
|
||||||
|
|
||||||
|
-- Monthly aggregates
|
||||||
|
INSERT INTO product_time_aggregates (product_id, period_type, period_start, quantity_sold, revenue)
|
||||||
|
SELECT
|
||||||
|
product_id,
|
||||||
|
'monthly' as period_type,
|
||||||
|
DATE(DATE_SUB(date, INTERVAL DAY(date)-1 DAY)) as period_start,
|
||||||
|
SUM(quantity) as quantity_sold,
|
||||||
|
SUM((price - COALESCE(discount, 0)) * quantity) as revenue
|
||||||
|
FROM orders
|
||||||
|
WHERE canceled = 0
|
||||||
|
GROUP BY product_id, DATE(DATE_SUB(date, INTERVAL DAY(date)-1 DAY));
|
||||||
|
`);
|
||||||
|
|
||||||
|
// Calculate vendor metrics
|
||||||
|
outputProgress({
|
||||||
|
status: 'running',
|
||||||
|
operation: 'Calculating vendor metrics',
|
||||||
|
percentage: '95'
|
||||||
|
});
|
||||||
|
|
||||||
|
await connection.query(`
|
||||||
|
INSERT INTO vendor_metrics (
|
||||||
|
vendor,
|
||||||
|
total_orders,
|
||||||
|
total_items_ordered,
|
||||||
|
total_items_received,
|
||||||
|
total_spend,
|
||||||
|
average_order_value,
|
||||||
|
fulfillment_rate,
|
||||||
|
average_delivery_days,
|
||||||
|
last_order_date,
|
||||||
|
last_delivery_date
|
||||||
|
)
|
||||||
|
SELECT
|
||||||
|
vendor,
|
||||||
|
COUNT(DISTINCT po_id) as total_orders,
|
||||||
|
SUM(ordered) as total_items_ordered,
|
||||||
|
SUM(received) as total_items_received,
|
||||||
|
SUM(cost_price * received) as total_spend,
|
||||||
|
AVG(cost_price * ordered) as average_order_value,
|
||||||
|
(SUM(received) / NULLIF(SUM(ordered), 0)) * 100 as fulfillment_rate,
|
||||||
|
AVG(DATEDIFF(received_date, date)) as average_delivery_days,
|
||||||
|
MAX(date) as last_order_date,
|
||||||
|
MAX(received_date) as last_delivery_date
|
||||||
|
FROM purchase_orders
|
||||||
|
WHERE status = 'closed'
|
||||||
|
GROUP BY vendor
|
||||||
|
ON DUPLICATE KEY UPDATE
|
||||||
|
total_orders = VALUES(total_orders),
|
||||||
|
total_items_ordered = VALUES(total_items_ordered),
|
||||||
|
total_items_received = VALUES(total_items_received),
|
||||||
|
total_spend = VALUES(total_spend),
|
||||||
|
average_order_value = VALUES(average_order_value),
|
||||||
|
fulfillment_rate = VALUES(fulfillment_rate),
|
||||||
|
average_delivery_days = VALUES(average_delivery_days),
|
||||||
|
last_order_date = VALUES(last_order_date),
|
||||||
|
last_delivery_date = VALUES(last_delivery_date);
|
||||||
|
`);
|
||||||
|
|
||||||
|
outputProgress({
|
||||||
|
status: 'complete',
|
||||||
|
operation: 'Metrics calculation completed',
|
||||||
|
percentage: '100'
|
||||||
|
});
|
||||||
|
|
||||||
|
} catch (error) {
|
||||||
|
logError(error, 'Error calculating metrics');
|
||||||
|
throw error;
|
||||||
|
} finally {
|
||||||
|
connection.release();
|
||||||
|
}
|
||||||
|
} catch (error) {
|
||||||
|
logError(error, 'Fatal error during metrics calculation');
|
||||||
|
throw error;
|
||||||
|
} finally {
|
||||||
|
if (pool) {
|
||||||
|
await pool.end();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Export the function if being required as a module
|
||||||
|
if (typeof module !== 'undefined' && module.exports) {
|
||||||
|
module.exports = calculateMetrics;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Run directly if called from command line
|
||||||
|
if (require.main === module) {
|
||||||
|
calculateMetrics().catch(error => {
|
||||||
|
console.error('Error:', error);
|
||||||
|
process.exit(1);
|
||||||
|
});
|
||||||
|
}
|
||||||
@@ -401,38 +401,75 @@ async function calculateVendorMetrics(connection) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Helper function to update product metrics
|
// Helper function to calculate metrics in batches
|
||||||
async function updateProductMetrics(connection, productId, startTime, current, total) {
|
async function calculateMetricsInBatch(connection) {
|
||||||
try {
|
try {
|
||||||
// Calculate sales velocity metrics
|
// Clear temporary tables
|
||||||
const velocityMetrics = await calculateSalesVelocity(connection, productId);
|
await connection.query('TRUNCATE TABLE temp_sales_metrics');
|
||||||
|
await connection.query('TRUNCATE TABLE temp_purchase_metrics');
|
||||||
|
|
||||||
// Calculate stock metrics
|
// Calculate sales metrics for all products in one go
|
||||||
const stockMetrics = await calculateStockMetrics(connection, productId, velocityMetrics.daily_sales_avg);
|
await connection.query(`
|
||||||
|
INSERT INTO temp_sales_metrics
|
||||||
|
SELECT
|
||||||
|
o.product_id,
|
||||||
|
COUNT(*) / NULLIF(DATEDIFF(MAX(o.date), MIN(o.date)), 0) as daily_sales_avg,
|
||||||
|
SUM(o.quantity) / NULLIF(DATEDIFF(MAX(o.date), MIN(o.date)), 0) * 7 as weekly_sales_avg,
|
||||||
|
SUM(o.quantity) / NULLIF(DATEDIFF(MAX(o.date), MIN(o.date)), 0) * 30 as monthly_sales_avg,
|
||||||
|
SUM(o.price * o.quantity) as total_revenue,
|
||||||
|
AVG((o.price - p.cost_price) / o.price * 100) as avg_margin_percent,
|
||||||
|
MIN(o.date) as first_sale_date,
|
||||||
|
MAX(o.date) as last_sale_date
|
||||||
|
FROM orders o
|
||||||
|
JOIN products p ON o.product_id = p.product_id
|
||||||
|
WHERE o.canceled = false
|
||||||
|
GROUP BY o.product_id
|
||||||
|
`);
|
||||||
|
|
||||||
// Calculate financial metrics
|
// Calculate purchase metrics for all products in one go
|
||||||
const financialMetrics = await calculateFinancialMetrics(connection, productId);
|
await connection.query(`
|
||||||
|
INSERT INTO temp_purchase_metrics
|
||||||
|
SELECT
|
||||||
|
product_id,
|
||||||
|
AVG(DATEDIFF(received_date, date)) as avg_lead_time_days,
|
||||||
|
MAX(date) as last_purchase_date,
|
||||||
|
MAX(received_date) as last_received_date
|
||||||
|
FROM purchase_orders
|
||||||
|
WHERE status = 'closed'
|
||||||
|
GROUP BY product_id
|
||||||
|
`);
|
||||||
|
|
||||||
// Calculate purchase metrics
|
// Update product_metrics table with all metrics at once
|
||||||
const purchaseMetrics = await calculatePurchaseMetrics(connection, productId);
|
|
||||||
|
|
||||||
// Update metrics in database
|
|
||||||
await connection.query(`
|
await connection.query(`
|
||||||
INSERT INTO product_metrics (
|
INSERT INTO product_metrics (
|
||||||
product_id,
|
product_id, daily_sales_avg, weekly_sales_avg, monthly_sales_avg,
|
||||||
daily_sales_avg,
|
days_of_inventory, weeks_of_inventory, safety_stock, reorder_point,
|
||||||
weekly_sales_avg,
|
avg_margin_percent, total_revenue, avg_lead_time_days,
|
||||||
monthly_sales_avg,
|
last_purchase_date, last_received_date
|
||||||
days_of_inventory,
|
)
|
||||||
weeks_of_inventory,
|
SELECT
|
||||||
safety_stock,
|
p.product_id,
|
||||||
reorder_point,
|
COALESCE(s.daily_sales_avg, 0),
|
||||||
total_revenue,
|
COALESCE(s.weekly_sales_avg, 0),
|
||||||
avg_margin_percent,
|
COALESCE(s.monthly_sales_avg, 0),
|
||||||
avg_lead_time_days,
|
CASE
|
||||||
last_purchase_date,
|
WHEN s.daily_sales_avg > 0 THEN FLOOR(p.stock_quantity / s.daily_sales_avg)
|
||||||
last_received_date
|
ELSE 999
|
||||||
) VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)
|
END as days_of_inventory,
|
||||||
|
CASE
|
||||||
|
WHEN s.daily_sales_avg > 0 THEN FLOOR(p.stock_quantity / s.daily_sales_avg / 7)
|
||||||
|
ELSE 999
|
||||||
|
END as weeks_of_inventory,
|
||||||
|
CEIL(COALESCE(s.daily_sales_avg, 0) * 14) as safety_stock,
|
||||||
|
CEIL(COALESCE(s.daily_sales_avg, 0) * 21) as reorder_point,
|
||||||
|
COALESCE(s.avg_margin_percent, 0),
|
||||||
|
COALESCE(s.total_revenue, 0),
|
||||||
|
COALESCE(pm.avg_lead_time_days, 0),
|
||||||
|
pm.last_purchase_date,
|
||||||
|
pm.last_received_date
|
||||||
|
FROM products p
|
||||||
|
LEFT JOIN temp_sales_metrics s ON p.product_id = s.product_id
|
||||||
|
LEFT JOIN temp_purchase_metrics pm ON p.product_id = pm.product_id
|
||||||
ON DUPLICATE KEY UPDATE
|
ON DUPLICATE KEY UPDATE
|
||||||
daily_sales_avg = VALUES(daily_sales_avg),
|
daily_sales_avg = VALUES(daily_sales_avg),
|
||||||
weekly_sales_avg = VALUES(weekly_sales_avg),
|
weekly_sales_avg = VALUES(weekly_sales_avg),
|
||||||
@@ -441,34 +478,37 @@ async function updateProductMetrics(connection, productId, startTime, current, t
|
|||||||
weeks_of_inventory = VALUES(weeks_of_inventory),
|
weeks_of_inventory = VALUES(weeks_of_inventory),
|
||||||
safety_stock = VALUES(safety_stock),
|
safety_stock = VALUES(safety_stock),
|
||||||
reorder_point = VALUES(reorder_point),
|
reorder_point = VALUES(reorder_point),
|
||||||
total_revenue = VALUES(total_revenue),
|
|
||||||
avg_margin_percent = VALUES(avg_margin_percent),
|
avg_margin_percent = VALUES(avg_margin_percent),
|
||||||
|
total_revenue = VALUES(total_revenue),
|
||||||
avg_lead_time_days = VALUES(avg_lead_time_days),
|
avg_lead_time_days = VALUES(avg_lead_time_days),
|
||||||
last_purchase_date = VALUES(last_purchase_date),
|
last_purchase_date = VALUES(last_purchase_date),
|
||||||
last_received_date = VALUES(last_received_date)
|
last_received_date = VALUES(last_received_date),
|
||||||
`, [
|
last_calculated_at = CURRENT_TIMESTAMP
|
||||||
productId,
|
`);
|
||||||
velocityMetrics.daily_sales_avg,
|
|
||||||
velocityMetrics.weekly_sales_avg,
|
// Calculate ABC classification in one go
|
||||||
velocityMetrics.monthly_sales_avg,
|
await connection.query(`
|
||||||
stockMetrics?.days_of_inventory || 0,
|
WITH revenue_ranks AS (
|
||||||
stockMetrics?.weeks_of_inventory || 0,
|
SELECT
|
||||||
stockMetrics?.safety_stock || 0,
|
product_id,
|
||||||
stockMetrics?.reorder_point || 0,
|
total_revenue,
|
||||||
financialMetrics.total_revenue,
|
total_revenue / SUM(total_revenue) OVER () * 100 as revenue_percent,
|
||||||
financialMetrics.avg_margin_percent,
|
ROW_NUMBER() OVER (ORDER BY total_revenue DESC) as rank
|
||||||
purchaseMetrics.avg_lead_time_days,
|
FROM product_metrics
|
||||||
purchaseMetrics.last_purchase_date,
|
WHERE total_revenue > 0
|
||||||
purchaseMetrics.last_received_date
|
)
|
||||||
]);
|
UPDATE product_metrics pm
|
||||||
|
JOIN revenue_ranks r ON pm.product_id = r.product_id
|
||||||
|
SET abc_class =
|
||||||
|
CASE
|
||||||
|
WHEN r.revenue_percent >= 20 THEN 'A'
|
||||||
|
WHEN r.revenue_percent >= 5 THEN 'B'
|
||||||
|
ELSE 'C'
|
||||||
|
END
|
||||||
|
`);
|
||||||
|
|
||||||
// Output progress every 5 products or every second
|
|
||||||
if (current % 5 === 0 || Date.now() - startTime > 1000) {
|
|
||||||
updateProgress(current, total, 'Calculating product metrics', startTime);
|
|
||||||
startTime = Date.now();
|
|
||||||
}
|
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logError(error, `Error updating metrics for product ${productId}`);
|
logError(error, 'Error in batch metrics calculation');
|
||||||
throw error;
|
throw error;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -1051,43 +1091,8 @@ async function main() {
|
|||||||
|
|
||||||
const connection = await pool.getConnection();
|
const connection = await pool.getConnection();
|
||||||
try {
|
try {
|
||||||
// Calculate product metrics
|
// Calculate metrics in batches
|
||||||
const [products] = await connection.query('SELECT DISTINCT product_id FROM products');
|
await calculateMetricsInBatch(connection);
|
||||||
const totalProducts = products.length;
|
|
||||||
let processedProducts = 0;
|
|
||||||
const metricsStartTime = Date.now();
|
|
||||||
|
|
||||||
outputProgress({
|
|
||||||
operation: 'Starting product metrics calculation',
|
|
||||||
message: `Calculating metrics for ${totalProducts} products...`,
|
|
||||||
current: 0,
|
|
||||||
total: totalProducts,
|
|
||||||
percentage: '0'
|
|
||||||
});
|
|
||||||
|
|
||||||
for (const product of products) {
|
|
||||||
try {
|
|
||||||
// Update progress every 5 products or 1 second
|
|
||||||
if (processedProducts % 5 === 0 || (Date.now() - lastUpdate) > 1000) {
|
|
||||||
updateProgress(processedProducts, totalProducts, 'Calculating product metrics', metricsStartTime);
|
|
||||||
lastUpdate = Date.now();
|
|
||||||
}
|
|
||||||
|
|
||||||
await updateProductMetrics(connection, product.product_id, metricsStartTime, processedProducts, totalProducts);
|
|
||||||
processedProducts++;
|
|
||||||
} catch (error) {
|
|
||||||
logError(error, `Error calculating metrics for product ${product.product_id}`);
|
|
||||||
// Continue with next product instead of failing completely
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
outputProgress({
|
|
||||||
operation: 'Product metrics calculation completed',
|
|
||||||
current: processedProducts,
|
|
||||||
total: totalProducts,
|
|
||||||
duration: formatDuration((Date.now() - metricsStartTime) / 1000),
|
|
||||||
percentage: '100'
|
|
||||||
});
|
|
||||||
|
|
||||||
// Calculate vendor metrics
|
// Calculate vendor metrics
|
||||||
await calculateVendorMetrics(connection);
|
await calculateVendorMetrics(connection);
|
||||||
|
|||||||
170
inventory-server/scripts/reset-metrics.js
Normal file
170
inventory-server/scripts/reset-metrics.js
Normal file
@@ -0,0 +1,170 @@
|
|||||||
|
const mysql = require('mysql2/promise');
|
||||||
|
const path = require('path');
|
||||||
|
const dotenv = require('dotenv');
|
||||||
|
const fs = require('fs');
|
||||||
|
|
||||||
|
dotenv.config({ path: path.join(__dirname, '../.env') });
|
||||||
|
|
||||||
|
const dbConfig = {
|
||||||
|
host: process.env.DB_HOST,
|
||||||
|
user: process.env.DB_USER,
|
||||||
|
password: process.env.DB_PASSWORD,
|
||||||
|
database: process.env.DB_NAME,
|
||||||
|
multipleStatements: true,
|
||||||
|
waitForConnections: true,
|
||||||
|
connectionLimit: 10,
|
||||||
|
queueLimit: 0,
|
||||||
|
namedPlaceholders: true
|
||||||
|
};
|
||||||
|
|
||||||
|
// Set up logging
|
||||||
|
const LOG_DIR = path.join(__dirname, '../logs');
|
||||||
|
const ERROR_LOG = path.join(LOG_DIR, 'import-errors.log');
|
||||||
|
const IMPORT_LOG = path.join(LOG_DIR, 'import.log');
|
||||||
|
|
||||||
|
// Ensure log directory exists
|
||||||
|
if (!fs.existsSync(LOG_DIR)) {
|
||||||
|
fs.mkdirSync(LOG_DIR, { recursive: true });
|
||||||
|
}
|
||||||
|
|
||||||
|
// Helper function to log errors
|
||||||
|
function logError(error, context = '') {
|
||||||
|
const timestamp = new Date().toISOString();
|
||||||
|
const errorMessage = `[${timestamp}] ${context}\nError: ${error.message}\nStack: ${error.stack}\n\n`;
|
||||||
|
fs.appendFileSync(ERROR_LOG, errorMessage);
|
||||||
|
console.error(`\n${context}\nError: ${error.message}`);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Helper function to log progress
|
||||||
|
function outputProgress(data) {
|
||||||
|
const timestamp = new Date().toISOString();
|
||||||
|
const logMessage = `[${timestamp}] ${JSON.stringify(data)}\n`;
|
||||||
|
fs.appendFileSync(IMPORT_LOG, logMessage);
|
||||||
|
console.log(JSON.stringify(data));
|
||||||
|
}
|
||||||
|
|
||||||
|
async function resetMetrics() {
|
||||||
|
let pool;
|
||||||
|
try {
|
||||||
|
pool = mysql.createPool(dbConfig);
|
||||||
|
const connection = await pool.getConnection();
|
||||||
|
|
||||||
|
try {
|
||||||
|
outputProgress({
|
||||||
|
status: 'running',
|
||||||
|
operation: 'Starting metrics reset',
|
||||||
|
message: 'Creating/resetting metrics tables...',
|
||||||
|
percentage: '0'
|
||||||
|
});
|
||||||
|
|
||||||
|
// Create tables if they don't exist and then truncate them
|
||||||
|
await connection.query(`
|
||||||
|
CREATE TABLE IF NOT EXISTS temp_sales_metrics (
|
||||||
|
product_id INT PRIMARY KEY,
|
||||||
|
total_quantity_sold INT DEFAULT 0,
|
||||||
|
total_revenue DECIMAL(10,2) DEFAULT 0.00,
|
||||||
|
average_price DECIMAL(10,2) DEFAULT 0.00,
|
||||||
|
last_sale_date DATE,
|
||||||
|
sales_rank INT
|
||||||
|
);
|
||||||
|
|
||||||
|
CREATE TABLE IF NOT EXISTS temp_purchase_metrics (
|
||||||
|
product_id INT PRIMARY KEY,
|
||||||
|
total_quantity_purchased INT DEFAULT 0,
|
||||||
|
total_cost DECIMAL(10,2) DEFAULT 0.00,
|
||||||
|
average_cost DECIMAL(10,2) DEFAULT 0.00,
|
||||||
|
last_purchase_date DATE,
|
||||||
|
purchase_rank INT
|
||||||
|
);
|
||||||
|
|
||||||
|
CREATE TABLE IF NOT EXISTS product_metrics (
|
||||||
|
product_id INT PRIMARY KEY,
|
||||||
|
total_quantity_sold INT DEFAULT 0,
|
||||||
|
total_revenue DECIMAL(10,2) DEFAULT 0.00,
|
||||||
|
average_price DECIMAL(10,2) DEFAULT 0.00,
|
||||||
|
total_quantity_purchased INT DEFAULT 0,
|
||||||
|
total_cost DECIMAL(10,2) DEFAULT 0.00,
|
||||||
|
average_cost DECIMAL(10,2) DEFAULT 0.00,
|
||||||
|
profit_margin DECIMAL(5,2) DEFAULT 0.00,
|
||||||
|
turnover_rate DECIMAL(5,2) DEFAULT 0.00,
|
||||||
|
abc_class CHAR(1),
|
||||||
|
last_calculated_at TIMESTAMP DEFAULT CURRENT_TIMESTAMP,
|
||||||
|
last_sale_date DATE,
|
||||||
|
last_purchase_date DATE,
|
||||||
|
sales_rank INT,
|
||||||
|
purchase_rank INT
|
||||||
|
);
|
||||||
|
|
||||||
|
CREATE TABLE IF NOT EXISTS product_time_aggregates (
|
||||||
|
product_id INT,
|
||||||
|
period_type ENUM('daily', 'weekly', 'monthly', 'quarterly', 'yearly'),
|
||||||
|
period_start DATE,
|
||||||
|
quantity_sold INT DEFAULT 0,
|
||||||
|
revenue DECIMAL(10,2) DEFAULT 0.00,
|
||||||
|
quantity_purchased INT DEFAULT 0,
|
||||||
|
purchase_cost DECIMAL(10,2) DEFAULT 0.00,
|
||||||
|
PRIMARY KEY (product_id, period_type, period_start)
|
||||||
|
);
|
||||||
|
|
||||||
|
CREATE TABLE IF NOT EXISTS vendor_metrics (
|
||||||
|
vendor VARCHAR(255) PRIMARY KEY,
|
||||||
|
total_orders INT DEFAULT 0,
|
||||||
|
total_items_ordered INT DEFAULT 0,
|
||||||
|
total_items_received INT DEFAULT 0,
|
||||||
|
total_spend DECIMAL(10,2) DEFAULT 0.00,
|
||||||
|
average_order_value DECIMAL(10,2) DEFAULT 0.00,
|
||||||
|
fulfillment_rate DECIMAL(5,2) DEFAULT 0.00,
|
||||||
|
average_delivery_days DECIMAL(5,1),
|
||||||
|
last_order_date DATE,
|
||||||
|
last_delivery_date DATE
|
||||||
|
);
|
||||||
|
|
||||||
|
TRUNCATE TABLE temp_sales_metrics;
|
||||||
|
TRUNCATE TABLE temp_purchase_metrics;
|
||||||
|
TRUNCATE TABLE product_metrics;
|
||||||
|
TRUNCATE TABLE product_time_aggregates;
|
||||||
|
TRUNCATE TABLE vendor_metrics;
|
||||||
|
`);
|
||||||
|
|
||||||
|
outputProgress({
|
||||||
|
status: 'complete',
|
||||||
|
operation: 'Metrics reset completed',
|
||||||
|
message: 'All metrics tables have been created/cleared',
|
||||||
|
percentage: '100'
|
||||||
|
});
|
||||||
|
|
||||||
|
} catch (error) {
|
||||||
|
logError(error, 'Error resetting metrics tables');
|
||||||
|
outputProgress({
|
||||||
|
status: 'error',
|
||||||
|
error: error.message,
|
||||||
|
operation: 'Failed to reset metrics'
|
||||||
|
});
|
||||||
|
throw error;
|
||||||
|
} finally {
|
||||||
|
connection.release();
|
||||||
|
}
|
||||||
|
} catch (error) {
|
||||||
|
logError(error, 'Fatal error during metrics reset');
|
||||||
|
outputProgress({
|
||||||
|
status: 'error',
|
||||||
|
error: error.message,
|
||||||
|
operation: 'Failed to reset metrics'
|
||||||
|
});
|
||||||
|
throw error;
|
||||||
|
} finally {
|
||||||
|
if (pool) {
|
||||||
|
await pool.end();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Run if called directly
|
||||||
|
if (require.main === module) {
|
||||||
|
resetMetrics().catch(error => {
|
||||||
|
logError(error, 'Unhandled error in main process');
|
||||||
|
process.exit(1);
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
module.exports = resetMetrics;
|
||||||
@@ -17,6 +17,7 @@ let importProgress = null;
|
|||||||
const updateClients = new Set();
|
const updateClients = new Set();
|
||||||
const importClients = new Set();
|
const importClients = new Set();
|
||||||
const resetClients = new Set();
|
const resetClients = new Set();
|
||||||
|
const resetMetricsClients = new Set();
|
||||||
|
|
||||||
// Helper to send progress to specific clients
|
// Helper to send progress to specific clients
|
||||||
function sendProgressToClients(clients, progress) {
|
function sendProgressToClients(clients, progress) {
|
||||||
@@ -107,6 +108,28 @@ router.get('/reset/progress', (req, res) => {
|
|||||||
});
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
|
// Add reset-metrics progress endpoint
|
||||||
|
router.get('/reset-metrics/progress', (req, res) => {
|
||||||
|
res.writeHead(200, {
|
||||||
|
'Content-Type': 'text/event-stream',
|
||||||
|
'Cache-Control': 'no-cache',
|
||||||
|
'Connection': 'keep-alive',
|
||||||
|
'Access-Control-Allow-Origin': req.headers.origin || '*',
|
||||||
|
'Access-Control-Allow-Credentials': 'true'
|
||||||
|
});
|
||||||
|
|
||||||
|
// Send an initial message to test the connection
|
||||||
|
res.write('data: {"status":"running","operation":"Initializing connection..."}\n\n');
|
||||||
|
|
||||||
|
// Add this client to the reset-metrics set
|
||||||
|
resetMetricsClients.add(res);
|
||||||
|
|
||||||
|
// Remove client when connection closes
|
||||||
|
req.on('close', () => {
|
||||||
|
resetMetricsClients.delete(res);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
// Debug endpoint to verify route registration
|
// Debug endpoint to verify route registration
|
||||||
router.get('/test', (req, res) => {
|
router.get('/test', (req, res) => {
|
||||||
console.log('CSV test endpoint hit');
|
console.log('CSV test endpoint hit');
|
||||||
@@ -434,4 +457,104 @@ router.post('/reset', async (req, res) => {
|
|||||||
}
|
}
|
||||||
});
|
});
|
||||||
|
|
||||||
|
// Add reset-metrics endpoint
|
||||||
|
router.post('/reset-metrics', async (req, res) => {
|
||||||
|
if (activeImport) {
|
||||||
|
res.status(400).json({ error: 'Operation already in progress' });
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
try {
|
||||||
|
// Set active import to prevent concurrent operations
|
||||||
|
activeImport = {
|
||||||
|
type: 'reset-metrics',
|
||||||
|
status: 'running',
|
||||||
|
operation: 'Starting metrics reset'
|
||||||
|
};
|
||||||
|
|
||||||
|
// Send initial response
|
||||||
|
res.status(200).json({ message: 'Reset metrics started' });
|
||||||
|
|
||||||
|
// Send initial progress through SSE
|
||||||
|
sendProgressToClients(resetMetricsClients, {
|
||||||
|
status: 'running',
|
||||||
|
operation: 'Starting metrics reset'
|
||||||
|
});
|
||||||
|
|
||||||
|
// Run the reset metrics script
|
||||||
|
const resetMetrics = require('../../scripts/reset-metrics');
|
||||||
|
await resetMetrics();
|
||||||
|
|
||||||
|
// Send completion through SSE
|
||||||
|
sendProgressToClients(resetMetricsClients, {
|
||||||
|
status: 'complete',
|
||||||
|
operation: 'Metrics reset completed'
|
||||||
|
});
|
||||||
|
|
||||||
|
activeImport = null;
|
||||||
|
} catch (error) {
|
||||||
|
console.error('Error during metrics reset:', error);
|
||||||
|
|
||||||
|
// Send error through SSE
|
||||||
|
sendProgressToClients(resetMetricsClients, {
|
||||||
|
status: 'error',
|
||||||
|
error: error.message || 'Failed to reset metrics'
|
||||||
|
});
|
||||||
|
|
||||||
|
activeImport = null;
|
||||||
|
res.status(500).json({ error: error.message || 'Failed to reset metrics' });
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
// Add calculate-metrics endpoint
|
||||||
|
router.post('/calculate-metrics', async (req, res) => {
|
||||||
|
if (activeImport) {
|
||||||
|
res.status(400).json({ error: 'Operation already in progress' });
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
try {
|
||||||
|
// Set active import to prevent concurrent operations
|
||||||
|
activeImport = {
|
||||||
|
type: 'calculate-metrics',
|
||||||
|
status: 'running',
|
||||||
|
operation: 'Starting metrics calculation'
|
||||||
|
};
|
||||||
|
|
||||||
|
// Send initial response
|
||||||
|
res.status(200).json({ message: 'Metrics calculation started' });
|
||||||
|
|
||||||
|
// Send initial progress through SSE
|
||||||
|
sendProgressToClients(importClients, {
|
||||||
|
status: 'running',
|
||||||
|
operation: 'Starting metrics calculation',
|
||||||
|
percentage: '0'
|
||||||
|
});
|
||||||
|
|
||||||
|
// Run the metrics calculation script
|
||||||
|
const calculateMetrics = require('../../scripts/calculate-metrics');
|
||||||
|
await calculateMetrics();
|
||||||
|
|
||||||
|
// Send completion through SSE
|
||||||
|
sendProgressToClients(importClients, {
|
||||||
|
status: 'complete',
|
||||||
|
operation: 'Metrics calculation completed',
|
||||||
|
percentage: '100'
|
||||||
|
});
|
||||||
|
|
||||||
|
activeImport = null;
|
||||||
|
} catch (error) {
|
||||||
|
console.error('Error during metrics calculation:', error);
|
||||||
|
|
||||||
|
// Send error through SSE
|
||||||
|
sendProgressToClients(importClients, {
|
||||||
|
status: 'error',
|
||||||
|
error: error.message || 'Failed to calculate metrics'
|
||||||
|
});
|
||||||
|
|
||||||
|
activeImport = null;
|
||||||
|
res.status(500).json({ error: error.message || 'Failed to calculate metrics' });
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
module.exports = router;
|
module.exports = router;
|
||||||
@@ -1,6 +1,8 @@
|
|||||||
|
const express = require('express');
|
||||||
|
const cors = require('cors');
|
||||||
|
const { spawn } = require('child_process');
|
||||||
const path = require('path');
|
const path = require('path');
|
||||||
const fs = require('fs');
|
const fs = require('fs');
|
||||||
const express = require('express');
|
|
||||||
const mysql = require('mysql2/promise');
|
const mysql = require('mysql2/promise');
|
||||||
const { corsMiddleware, corsErrorHandler } = require('./middleware/cors');
|
const { corsMiddleware, corsErrorHandler } = require('./middleware/cors');
|
||||||
const { initPool } = require('./utils/db');
|
const { initPool } = require('./utils/db');
|
||||||
@@ -127,6 +129,111 @@ pool.getConnection()
|
|||||||
process.exit(1);
|
process.exit(1);
|
||||||
});
|
});
|
||||||
|
|
||||||
|
// Initialize client sets for SSE
|
||||||
|
const importClients = new Set();
|
||||||
|
const updateClients = new Set();
|
||||||
|
const resetClients = new Set();
|
||||||
|
const resetMetricsClients = new Set();
|
||||||
|
|
||||||
|
// Helper function to send progress to SSE clients
|
||||||
|
const sendProgressToClients = (clients, data) => {
|
||||||
|
clients.forEach(client => {
|
||||||
|
try {
|
||||||
|
client.write(`data: ${JSON.stringify(data)}\n\n`);
|
||||||
|
} catch (error) {
|
||||||
|
console.error('Error sending SSE update:', error);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
};
|
||||||
|
|
||||||
|
// Setup SSE connection
|
||||||
|
const setupSSE = (req, res) => {
|
||||||
|
const { type } = req.params;
|
||||||
|
|
||||||
|
// Set headers for SSE
|
||||||
|
res.writeHead(200, {
|
||||||
|
'Content-Type': 'text/event-stream',
|
||||||
|
'Cache-Control': 'no-cache',
|
||||||
|
'Connection': 'keep-alive',
|
||||||
|
'Access-Control-Allow-Origin': req.headers.origin || '*',
|
||||||
|
'Access-Control-Allow-Credentials': 'true'
|
||||||
|
});
|
||||||
|
|
||||||
|
// Send initial message
|
||||||
|
res.write('data: {"status":"connected"}\n\n');
|
||||||
|
|
||||||
|
// Add client to appropriate set
|
||||||
|
const clientSet = type === 'import' ? importClients :
|
||||||
|
type === 'update' ? updateClients :
|
||||||
|
type === 'reset' ? resetClients :
|
||||||
|
type === 'reset-metrics' ? resetMetricsClients :
|
||||||
|
null;
|
||||||
|
|
||||||
|
if (clientSet) {
|
||||||
|
clientSet.add(res);
|
||||||
|
|
||||||
|
// Remove client when connection closes
|
||||||
|
req.on('close', () => {
|
||||||
|
clientSet.delete(res);
|
||||||
|
});
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
// Update the status endpoint to include reset-metrics
|
||||||
|
app.get('/csv/status', (req, res) => {
|
||||||
|
res.json({
|
||||||
|
active: !!currentOperation,
|
||||||
|
type: currentOperation?.type || null,
|
||||||
|
progress: currentOperation ? {
|
||||||
|
status: currentOperation.status,
|
||||||
|
operation: currentOperation.operation,
|
||||||
|
current: currentOperation.current,
|
||||||
|
total: currentOperation.total,
|
||||||
|
percentage: currentOperation.percentage
|
||||||
|
} : null
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
// Update progress endpoint mapping
|
||||||
|
app.get('/csv/:type/progress', (req, res) => {
|
||||||
|
const { type } = req.params;
|
||||||
|
if (!['import', 'update', 'reset', 'reset-metrics'].includes(type)) {
|
||||||
|
res.status(400).json({ error: 'Invalid operation type' });
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
setupSSE(req, res);
|
||||||
|
});
|
||||||
|
|
||||||
|
// Update the cancel endpoint to handle reset-metrics
|
||||||
|
app.post('/csv/cancel', (req, res) => {
|
||||||
|
const { operation } = req.query;
|
||||||
|
|
||||||
|
if (!currentOperation) {
|
||||||
|
res.status(400).json({ error: 'No operation in progress' });
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (operation && operation.toLowerCase() !== currentOperation.type) {
|
||||||
|
res.status(400).json({ error: 'Operation type mismatch' });
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
try {
|
||||||
|
// Handle cancellation based on operation type
|
||||||
|
if (currentOperation.type === 'reset-metrics') {
|
||||||
|
// Reset metrics doesn't need special cleanup
|
||||||
|
currentOperation = null;
|
||||||
|
res.json({ message: 'Reset metrics cancelled' });
|
||||||
|
} else {
|
||||||
|
// ... existing cancellation logic for other operations ...
|
||||||
|
}
|
||||||
|
} catch (error) {
|
||||||
|
console.error('Error during cancellation:', error);
|
||||||
|
res.status(500).json({ error: 'Failed to cancel operation' });
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
const PORT = process.env.PORT || 3000;
|
const PORT = process.env.PORT || 3000;
|
||||||
app.listen(PORT, () => {
|
app.listen(PORT, () => {
|
||||||
console.log(`[Server] Running in ${process.env.NODE_ENV || 'development'} mode on port ${PORT}`);
|
console.log(`[Server] Running in ${process.env.NODE_ENV || 'development'} mode on port ${PORT}`);
|
||||||
|
|||||||
@@ -58,9 +58,10 @@ export function Settings() {
|
|||||||
orders: 0,
|
orders: 0,
|
||||||
purchaseOrders: 0
|
purchaseOrders: 0
|
||||||
});
|
});
|
||||||
const [isCreatingSnapshot, setIsCreatingSnapshot] = useState(false);
|
const [isResettingMetrics, setIsResettingMetrics] = useState(false);
|
||||||
const [isRestoringSnapshot, setIsRestoringSnapshot] = useState(false);
|
const [resetMetricsProgress, setResetMetricsProgress] = useState<ImportProgress | null>(null);
|
||||||
const [snapshotProgress, setSnapshotProgress] = useState<ImportProgress | null>(null);
|
const [isCalculatingMetrics, setIsCalculatingMetrics] = useState(false);
|
||||||
|
const [metricsProgress, setMetricsProgress] = useState<ImportProgress | null>(null);
|
||||||
|
|
||||||
// Helper function to update progress state
|
// Helper function to update progress state
|
||||||
const updateProgressState = (progressData: any) => {
|
const updateProgressState = (progressData: any) => {
|
||||||
@@ -92,13 +93,11 @@ export function Settings() {
|
|||||||
setPurchaseOrdersProgress(prev => ({ ...prev, ...progressUpdate }));
|
setPurchaseOrdersProgress(prev => ({ ...prev, ...progressUpdate }));
|
||||||
} else if (operation.includes('metrics') || operation.includes('vendor metrics')) {
|
} else if (operation.includes('metrics') || operation.includes('vendor metrics')) {
|
||||||
setImportProgress(prev => ({ ...prev, ...progressUpdate }));
|
setImportProgress(prev => ({ ...prev, ...progressUpdate }));
|
||||||
} else if (operation.includes('snapshot')) {
|
|
||||||
setSnapshotProgress(prev => ({ ...prev, ...progressUpdate }));
|
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
// Helper to connect to event source
|
// Helper to connect to event source
|
||||||
const connectToEventSource = useCallback((type: 'update' | 'import' | 'reset' | 'snapshot') => {
|
const connectToEventSource = useCallback((type: 'update' | 'import' | 'reset' | 'reset-metrics' | 'calculate-metrics') => {
|
||||||
if (eventSource) {
|
if (eventSource) {
|
||||||
eventSource.close();
|
eventSource.close();
|
||||||
}
|
}
|
||||||
@@ -122,6 +121,8 @@ export function Settings() {
|
|||||||
// For non-import operations, use the existing logic
|
// For non-import operations, use the existing logic
|
||||||
const setProgress = type === 'update' ? setUpdateProgress :
|
const setProgress = type === 'update' ? setUpdateProgress :
|
||||||
type === 'reset' ? setResetProgress :
|
type === 'reset' ? setResetProgress :
|
||||||
|
type === 'reset-metrics' ? setResetMetricsProgress :
|
||||||
|
type === 'calculate-metrics' ? setMetricsProgress :
|
||||||
setImportProgress;
|
setImportProgress;
|
||||||
setProgress(prev => ({
|
setProgress(prev => ({
|
||||||
...prev,
|
...prev,
|
||||||
@@ -147,6 +148,8 @@ export function Settings() {
|
|||||||
if (type === 'update') setIsUpdating(true);
|
if (type === 'update') setIsUpdating(true);
|
||||||
else if (type === 'import') setIsImporting(true);
|
else if (type === 'import') setIsImporting(true);
|
||||||
else if (type === 'reset') setIsResetting(true);
|
else if (type === 'reset') setIsResetting(true);
|
||||||
|
else if (type === 'reset-metrics') setIsResettingMetrics(true);
|
||||||
|
else if (type === 'calculate-metrics') setIsCalculatingMetrics(true);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (progressData.status === 'complete') {
|
if (progressData.status === 'complete') {
|
||||||
@@ -164,6 +167,12 @@ export function Settings() {
|
|||||||
} else if (type === 'reset') {
|
} else if (type === 'reset') {
|
||||||
setIsResetting(false);
|
setIsResetting(false);
|
||||||
setResetProgress(null);
|
setResetProgress(null);
|
||||||
|
} else if (type === 'reset-metrics') {
|
||||||
|
setIsResettingMetrics(false);
|
||||||
|
setResetMetricsProgress(null);
|
||||||
|
} else if (type === 'calculate-metrics') {
|
||||||
|
setIsCalculatingMetrics(false);
|
||||||
|
setMetricsProgress(null);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!progressData.operation?.includes('cancelled')) {
|
if (!progressData.operation?.includes('cancelled')) {
|
||||||
@@ -180,6 +189,10 @@ export function Settings() {
|
|||||||
setIsImporting(false);
|
setIsImporting(false);
|
||||||
} else if (type === 'reset') {
|
} else if (type === 'reset') {
|
||||||
setIsResetting(false);
|
setIsResetting(false);
|
||||||
|
} else if (type === 'reset-metrics') {
|
||||||
|
setIsResettingMetrics(false);
|
||||||
|
} else if (type === 'calculate-metrics') {
|
||||||
|
setIsCalculatingMetrics(false);
|
||||||
}
|
}
|
||||||
|
|
||||||
handleError(`${type.charAt(0).toUpperCase() + type.slice(1)}`, progressData.error || 'Unknown error');
|
handleError(`${type.charAt(0).toUpperCase() + type.slice(1)}`, progressData.error || 'Unknown error');
|
||||||
@@ -188,7 +201,7 @@ export function Settings() {
|
|||||||
console.error('Error parsing event data:', error);
|
console.error('Error parsing event data:', error);
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
}, []); // Remove dependencies that might prevent initial connection
|
}, []);
|
||||||
|
|
||||||
// Check for active operations on mount
|
// Check for active operations on mount
|
||||||
useEffect(() => {
|
useEffect(() => {
|
||||||
@@ -203,13 +216,15 @@ export function Settings() {
|
|||||||
|
|
||||||
if (data.active) {
|
if (data.active) {
|
||||||
// Try to determine the operation type from progress if available
|
// Try to determine the operation type from progress if available
|
||||||
let operationType: 'update' | 'import' | 'reset' | null = null;
|
let operationType: 'update' | 'import' | 'reset' | 'reset-metrics' | 'calculate-metrics' | null = null;
|
||||||
|
|
||||||
if (data.progress?.operation) {
|
if (data.progress?.operation) {
|
||||||
const operation = data.progress.operation.toLowerCase();
|
const operation = data.progress.operation.toLowerCase();
|
||||||
if (operation.includes('update')) operationType = 'update';
|
if (operation.includes('update')) operationType = 'update';
|
||||||
else if (operation.includes('import')) operationType = 'import';
|
else if (operation.includes('import')) operationType = 'import';
|
||||||
else if (operation.includes('reset')) operationType = 'reset';
|
else if (operation.includes('reset')) operationType = 'reset';
|
||||||
|
else if (operation.includes('reset-metrics')) operationType = 'reset-metrics';
|
||||||
|
else if (operation.includes('calculate-metrics')) operationType = 'calculate-metrics';
|
||||||
} else {
|
} else {
|
||||||
// If no progress data, try to connect to import stream by default
|
// If no progress data, try to connect to import stream by default
|
||||||
// since that's the most common long-running operation
|
// since that's the most common long-running operation
|
||||||
@@ -242,6 +257,22 @@ export function Settings() {
|
|||||||
status: data.progress.status || 'running'
|
status: data.progress.status || 'running'
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
} else if (operationType === 'reset-metrics') {
|
||||||
|
setIsResettingMetrics(true);
|
||||||
|
if (data.progress) {
|
||||||
|
setResetMetricsProgress({
|
||||||
|
...data.progress,
|
||||||
|
status: data.progress.status || 'running'
|
||||||
|
});
|
||||||
|
}
|
||||||
|
} else if (operationType === 'calculate-metrics') {
|
||||||
|
setIsCalculatingMetrics(true);
|
||||||
|
if (data.progress) {
|
||||||
|
setMetricsProgress({
|
||||||
|
...data.progress,
|
||||||
|
status: data.progress.status || 'running'
|
||||||
|
});
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Connect to the appropriate event source
|
// Connect to the appropriate event source
|
||||||
@@ -258,6 +289,14 @@ export function Settings() {
|
|||||||
}, []); // Remove connectToEventSource dependency to ensure it runs on mount
|
}, []); // Remove connectToEventSource dependency to ensure it runs on mount
|
||||||
|
|
||||||
// Clean up function to reset state
|
// Clean up function to reset state
|
||||||
|
useEffect(() => {
|
||||||
|
return () => {
|
||||||
|
if (eventSource) {
|
||||||
|
console.log('Cleaning up event source'); // Debug log
|
||||||
|
eventSource.close();
|
||||||
|
}
|
||||||
|
};
|
||||||
|
}, [eventSource]);
|
||||||
|
|
||||||
const handleCancel = async () => {
|
const handleCancel = async () => {
|
||||||
// Determine which operation is running first
|
// Determine which operation is running first
|
||||||
@@ -396,72 +435,83 @@ export function Settings() {
|
|||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
// Add handlers for snapshot operations
|
const handleResetMetrics = async () => {
|
||||||
const handleCreateSnapshot = async () => {
|
setIsResettingMetrics(true);
|
||||||
|
setResetMetricsProgress({ status: 'running', operation: 'Starting metrics reset' });
|
||||||
|
|
||||||
try {
|
try {
|
||||||
setIsCreatingSnapshot(true);
|
|
||||||
setSnapshotProgress({ status: 'running', operation: 'Creating test data snapshot' });
|
|
||||||
|
|
||||||
// Connect to SSE for progress updates
|
// Connect to SSE for progress updates
|
||||||
connectToEventSource('snapshot');
|
connectToEventSource('reset-metrics');
|
||||||
|
|
||||||
const response = await fetch(`${config.apiUrl}/snapshot/create`, {
|
// Make the reset request
|
||||||
|
const response = await fetch(`${config.apiUrl}/csv/reset-metrics`, {
|
||||||
method: 'POST',
|
method: 'POST',
|
||||||
credentials: 'include'
|
credentials: 'include'
|
||||||
});
|
});
|
||||||
|
|
||||||
if (!response.ok) {
|
if (!response.ok) {
|
||||||
throw new Error('Failed to create snapshot');
|
const data = await response.json().catch(() => ({}));
|
||||||
|
throw new Error(data.error || 'Failed to reset metrics');
|
||||||
}
|
}
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
console.error('Error creating snapshot:', error);
|
|
||||||
if (eventSource) {
|
if (eventSource) {
|
||||||
eventSource.close();
|
eventSource.close();
|
||||||
setEventSource(null);
|
setEventSource(null);
|
||||||
}
|
}
|
||||||
setIsCreatingSnapshot(false);
|
setIsResettingMetrics(false);
|
||||||
setSnapshotProgress(null);
|
setResetMetricsProgress(null);
|
||||||
toast.error('Failed to create snapshot');
|
handleError('Metrics reset', error instanceof Error ? error.message : 'Unknown error');
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
const handleRestoreSnapshot = async () => {
|
const handleCalculateMetrics = async () => {
|
||||||
|
setIsCalculatingMetrics(true);
|
||||||
|
setMetricsProgress({ status: 'running', operation: 'Starting metrics calculation' });
|
||||||
|
|
||||||
try {
|
try {
|
||||||
setIsRestoringSnapshot(true);
|
|
||||||
setSnapshotProgress({ status: 'running', operation: 'Restoring test data snapshot' });
|
|
||||||
|
|
||||||
// Connect to SSE for progress updates
|
// Connect to SSE for progress updates
|
||||||
connectToEventSource('snapshot');
|
connectToEventSource('calculate-metrics');
|
||||||
|
|
||||||
const response = await fetch(`${config.apiUrl}/snapshot/restore`, {
|
// Make the calculation request
|
||||||
|
const response = await fetch(`${config.apiUrl}/csv/calculate-metrics`, {
|
||||||
method: 'POST',
|
method: 'POST',
|
||||||
credentials: 'include'
|
credentials: 'include'
|
||||||
});
|
});
|
||||||
|
|
||||||
if (!response.ok) {
|
if (!response.ok) {
|
||||||
throw new Error('Failed to restore snapshot');
|
const data = await response.json().catch(() => ({}));
|
||||||
|
throw new Error(data.error || 'Failed to calculate metrics');
|
||||||
}
|
}
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
console.error('Error restoring snapshot:', error);
|
|
||||||
if (eventSource) {
|
if (eventSource) {
|
||||||
eventSource.close();
|
eventSource.close();
|
||||||
setEventSource(null);
|
setEventSource(null);
|
||||||
}
|
}
|
||||||
setIsRestoringSnapshot(false);
|
setIsCalculatingMetrics(false);
|
||||||
setSnapshotProgress(null);
|
setMetricsProgress(null);
|
||||||
toast.error('Failed to restore snapshot');
|
handleError('Metrics calculation', error instanceof Error ? error.message : 'Unknown error');
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
// Cleanup on unmount
|
// Update the message handlers to use toast
|
||||||
useEffect(() => {
|
const handleComplete = (operation: string) => {
|
||||||
return () => {
|
toast.success(`${operation} completed successfully`);
|
||||||
if (eventSource) {
|
};
|
||||||
console.log('Cleaning up event source'); // Debug log
|
|
||||||
eventSource.close();
|
const handleError = (operation: string, error: string) => {
|
||||||
}
|
// Skip error toast if we're cancelling or if it's a cancellation error
|
||||||
|
if (error.includes('cancelled') ||
|
||||||
|
error.includes('Process exited with code 143') ||
|
||||||
|
error.includes('Operation cancelled') ||
|
||||||
|
error.includes('500 Internal Server Error') ||
|
||||||
|
// Skip "Failed to start" errors if we have active progress
|
||||||
|
(error.includes('Failed to start CSV import') && (importProgress || purchaseOrdersProgress)) ||
|
||||||
|
// Skip connection errors if we have active progress
|
||||||
|
(error.includes('Failed to fetch') && (importProgress || purchaseOrdersProgress))) {
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
toast.error(`${operation} failed: ${error}`);
|
||||||
};
|
};
|
||||||
}, [eventSource]);
|
|
||||||
|
|
||||||
const renderProgress = (progress: ImportProgress | null) => {
|
const renderProgress = (progress: ImportProgress | null) => {
|
||||||
if (!progress) return null;
|
if (!progress) return null;
|
||||||
@@ -523,26 +573,6 @@ export function Settings() {
|
|||||||
);
|
);
|
||||||
};
|
};
|
||||||
|
|
||||||
const handleError = (operation: string, error: string) => {
|
|
||||||
// Skip error toast if we're cancelling or if it's a cancellation error
|
|
||||||
if (error.includes('cancelled') ||
|
|
||||||
error.includes('Process exited with code 143') ||
|
|
||||||
error.includes('Operation cancelled') ||
|
|
||||||
error.includes('500 Internal Server Error') ||
|
|
||||||
// Skip "Failed to start" errors if we have active progress
|
|
||||||
(error.includes('Failed to start CSV import') && (importProgress || purchaseOrdersProgress)) ||
|
|
||||||
// Skip connection errors if we have active progress
|
|
||||||
(error.includes('Failed to fetch') && (importProgress || purchaseOrdersProgress))) {
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
toast.error(`${operation} failed: ${error}`);
|
|
||||||
};
|
|
||||||
|
|
||||||
// Update the message handlers to use toast
|
|
||||||
const handleComplete = (operation: string) => {
|
|
||||||
toast.success(`${operation} completed successfully`);
|
|
||||||
};
|
|
||||||
|
|
||||||
return (
|
return (
|
||||||
<div className="p-8 space-y-8">
|
<div className="p-8 space-y-8">
|
||||||
<div className="flex items-center justify-between">
|
<div className="flex items-center justify-between">
|
||||||
@@ -718,16 +748,20 @@ export function Settings() {
|
|||||||
</CardContent>
|
</CardContent>
|
||||||
</Card>
|
</Card>
|
||||||
|
|
||||||
{/* Reset Database Card */}
|
{/* Database Management Card */}
|
||||||
<Card>
|
<Card>
|
||||||
<CardHeader>
|
<CardHeader>
|
||||||
<CardTitle>Reset Database</CardTitle>
|
<CardTitle>Database Management</CardTitle>
|
||||||
<CardDescription>Drop all tables and recreate the database schema. This will delete ALL data.</CardDescription>
|
<CardDescription>Reset database or metrics tables</CardDescription>
|
||||||
</CardHeader>
|
</CardHeader>
|
||||||
<CardContent>
|
<CardContent className="space-y-4">
|
||||||
|
<div className="flex gap-4">
|
||||||
<AlertDialog>
|
<AlertDialog>
|
||||||
<AlertDialogTrigger asChild>
|
<AlertDialogTrigger asChild>
|
||||||
<Button variant="destructive" disabled={isResetting || isImporting || isUpdating}>
|
<Button
|
||||||
|
variant="destructive"
|
||||||
|
disabled={isResetting || isImporting || isUpdating || isResettingMetrics}
|
||||||
|
>
|
||||||
Reset Database
|
Reset Database
|
||||||
</Button>
|
</Button>
|
||||||
</AlertDialogTrigger>
|
</AlertDialogTrigger>
|
||||||
@@ -744,6 +778,32 @@ export function Settings() {
|
|||||||
</AlertDialogFooter>
|
</AlertDialogFooter>
|
||||||
</AlertDialogContent>
|
</AlertDialogContent>
|
||||||
</AlertDialog>
|
</AlertDialog>
|
||||||
|
|
||||||
|
<AlertDialog>
|
||||||
|
<AlertDialogTrigger asChild>
|
||||||
|
<Button
|
||||||
|
variant="secondary"
|
||||||
|
disabled={isResetting || isImporting || isUpdating || isResettingMetrics}
|
||||||
|
>
|
||||||
|
Reset Metrics Only
|
||||||
|
</Button>
|
||||||
|
</AlertDialogTrigger>
|
||||||
|
<AlertDialogContent>
|
||||||
|
<AlertDialogHeader>
|
||||||
|
<AlertDialogTitle>Reset metrics tables?</AlertDialogTitle>
|
||||||
|
<AlertDialogDescription>
|
||||||
|
This will clear all metrics tables while preserving your core data (products, orders, etc.).
|
||||||
|
You can then recalculate metrics with the Import Data function.
|
||||||
|
</AlertDialogDescription>
|
||||||
|
</AlertDialogHeader>
|
||||||
|
<AlertDialogFooter>
|
||||||
|
<AlertDialogCancel>Cancel</AlertDialogCancel>
|
||||||
|
<AlertDialogAction onClick={handleResetMetrics}>Continue</AlertDialogAction>
|
||||||
|
</AlertDialogFooter>
|
||||||
|
</AlertDialogContent>
|
||||||
|
</AlertDialog>
|
||||||
|
</div>
|
||||||
|
|
||||||
{resetProgress && (
|
{resetProgress && (
|
||||||
<div className="mt-4">
|
<div className="mt-4">
|
||||||
<Progress value={Number(resetProgress.percentage)} className="mb-2" />
|
<Progress value={Number(resetProgress.percentage)} className="mb-2" />
|
||||||
@@ -752,84 +812,63 @@ export function Settings() {
|
|||||||
</p>
|
</p>
|
||||||
</div>
|
</div>
|
||||||
)}
|
)}
|
||||||
</CardContent>
|
|
||||||
</Card>
|
|
||||||
|
|
||||||
{/* Test Data Snapshots Card */}
|
{resetMetricsProgress && (
|
||||||
<Card>
|
<div className="mt-4">
|
||||||
<CardHeader>
|
<Progress value={Number(resetMetricsProgress.percentage)} className="mb-2" />
|
||||||
<CardTitle>Test Data Snapshots</CardTitle>
|
|
||||||
<CardDescription>Create and restore test data snapshots for development and testing.</CardDescription>
|
|
||||||
</CardHeader>
|
|
||||||
<CardContent className="space-y-4">
|
|
||||||
<div className="flex space-x-4">
|
|
||||||
<Button
|
|
||||||
onClick={handleCreateSnapshot}
|
|
||||||
disabled={isCreatingSnapshot || isRestoringSnapshot || isImporting || isUpdating || isResetting}
|
|
||||||
>
|
|
||||||
{isCreatingSnapshot ? (
|
|
||||||
<>
|
|
||||||
<Loader2 className="mr-2 h-4 w-4 animate-spin" />
|
|
||||||
Creating Snapshot...
|
|
||||||
</>
|
|
||||||
) : (
|
|
||||||
<>Create Snapshot</>
|
|
||||||
)}
|
|
||||||
</Button>
|
|
||||||
<AlertDialog>
|
|
||||||
<AlertDialogTrigger asChild>
|
|
||||||
<Button
|
|
||||||
variant="secondary"
|
|
||||||
disabled={isCreatingSnapshot || isRestoringSnapshot || isImporting || isUpdating || isResetting}
|
|
||||||
>
|
|
||||||
{isRestoringSnapshot ? (
|
|
||||||
<>
|
|
||||||
<Loader2 className="mr-2 h-4 w-4 animate-spin" />
|
|
||||||
Restoring...
|
|
||||||
</>
|
|
||||||
) : (
|
|
||||||
<>Restore Snapshot</>
|
|
||||||
)}
|
|
||||||
</Button>
|
|
||||||
</AlertDialogTrigger>
|
|
||||||
<AlertDialogContent>
|
|
||||||
<AlertDialogHeader>
|
|
||||||
<AlertDialogTitle>Restore test data snapshot?</AlertDialogTitle>
|
|
||||||
<AlertDialogDescription>
|
|
||||||
This will replace your current database with the test data snapshot. Any unsaved changes will be lost.
|
|
||||||
</AlertDialogDescription>
|
|
||||||
</AlertDialogHeader>
|
|
||||||
<AlertDialogFooter>
|
|
||||||
<AlertDialogCancel>Cancel</AlertDialogCancel>
|
|
||||||
<AlertDialogAction onClick={handleRestoreSnapshot}>Continue</AlertDialogAction>
|
|
||||||
</AlertDialogFooter>
|
|
||||||
</AlertDialogContent>
|
|
||||||
</AlertDialog>
|
|
||||||
</div>
|
|
||||||
{snapshotProgress && (
|
|
||||||
<div>
|
|
||||||
<Progress value={Number(snapshotProgress.percentage)} className="mb-2" />
|
|
||||||
<p className="text-sm text-muted-foreground">
|
<p className="text-sm text-muted-foreground">
|
||||||
{snapshotProgress.message || 'Processing snapshot...'}
|
{resetMetricsProgress.message || 'Resetting metrics...'}
|
||||||
</p>
|
</p>
|
||||||
</div>
|
</div>
|
||||||
)}
|
)}
|
||||||
<div className="text-sm text-muted-foreground">
|
|
||||||
<p>The test data snapshot includes:</p>
|
|
||||||
<ul className="list-disc list-inside mt-2">
|
|
||||||
<li>~100 diverse products with associated data</li>
|
|
||||||
<li>Orders from the last 6 months</li>
|
|
||||||
<li>Purchase orders from the last 6 months</li>
|
|
||||||
<li>Categories and product relationships</li>
|
|
||||||
</ul>
|
|
||||||
</div>
|
|
||||||
</CardContent>
|
</CardContent>
|
||||||
</Card>
|
</Card>
|
||||||
|
|
||||||
{/* Show progress outside cards if neither operation is running but we have progress state */}
|
{/* Add new Metrics Calculation Card */}
|
||||||
{!isUpdating && !isImporting && !isResetting && (updateProgress || importProgress || resetProgress) && (
|
<Card>
|
||||||
|
<CardHeader>
|
||||||
|
<CardTitle>Metrics Calculation</CardTitle>
|
||||||
|
<CardDescription>Calculate metrics for all products based on current data</CardDescription>
|
||||||
|
</CardHeader>
|
||||||
|
<CardContent>
|
||||||
|
<div className="flex gap-2">
|
||||||
|
<Button
|
||||||
|
className="flex-1"
|
||||||
|
onClick={handleCalculateMetrics}
|
||||||
|
disabled={isCalculatingMetrics || isImporting || isUpdating || isResetting || isResettingMetrics}
|
||||||
|
>
|
||||||
|
{isCalculatingMetrics ? (
|
||||||
<>
|
<>
|
||||||
{renderProgress(updateProgress || importProgress || resetProgress)}
|
<Loader2 className="mr-2 h-4 w-4 animate-spin" />
|
||||||
|
Calculating Metrics...
|
||||||
|
</>
|
||||||
|
) : (
|
||||||
|
<>
|
||||||
|
<RefreshCw className="mr-2 h-4 w-4" />
|
||||||
|
Calculate Metrics
|
||||||
|
</>
|
||||||
|
)}
|
||||||
|
</Button>
|
||||||
|
|
||||||
|
{isCalculatingMetrics && (
|
||||||
|
<Button
|
||||||
|
variant="destructive"
|
||||||
|
onClick={handleCancel}
|
||||||
|
>
|
||||||
|
<X className="h-4 w-4" />
|
||||||
|
</Button>
|
||||||
|
)}
|
||||||
|
</div>
|
||||||
|
|
||||||
|
{metricsProgress && renderProgress(metricsProgress)}
|
||||||
|
</CardContent>
|
||||||
|
</Card>
|
||||||
|
|
||||||
|
{/* Show progress outside cards if no operation is running but we have progress state */}
|
||||||
|
{!isUpdating && !isImporting && !isResetting && !isResettingMetrics && !isCalculatingMetrics &&
|
||||||
|
(updateProgress || importProgress || resetProgress || resetMetricsProgress || metricsProgress) && (
|
||||||
|
<>
|
||||||
|
{renderProgress(updateProgress || importProgress || resetProgress || resetMetricsProgress || metricsProgress)}
|
||||||
</>
|
</>
|
||||||
)}
|
)}
|
||||||
</div>
|
</div>
|
||||||
|
|||||||
Reference in New Issue
Block a user