import os import re import sys import time import pandas as pd import matplotlib.pyplot as plt from datetime import datetime from matplotlib.lines import Line2D from typing import Optional, Tuple, List, Dict, Any, Union from pathlib import Path import numpy as np import base64 from io import BytesIO from jinja2 import Template from colorama import Fore, Style, init import multiprocessing as mp from concurrent.futures import ThreadPoolExecutor, ProcessPoolExecutor import psutil # 初始化colorama init(autoreset=True) # 避免 SettingWithCopy 警告影响输出可读性 pd.options.mode.chained_assignment = None # 设置中文字体支持 plt.rcParams['font.sans-serif'] = ['SimHei', 'DejaVu Sans', 'Arial Unicode MS', 'Microsoft YaHei'] plt.rcParams['axes.unicode_minus'] = False # HTML模板 - 更新为支持多关键词 # 性能优化配置 OPTIMIZATION_CONFIG = { 'max_workers': min(mp.cpu_count(), 8), # 限制最大工作线程数 'chunk_size': 50000, # 分块读取大小 'use_threading': True, # 使用多线程 'memory_limit_gb': psutil.virtual_memory().available // (1024 ** 3) * 0.7, # 内存限制 } HTML_TEMPLATE = """ 测试报告分析 - {{ keywords_display }}

📊 测试报告分析 (多文件合并)

关键词: {% for keyword in keywords %} {{ keyword }} {% endfor %}

生成时间: {{ timestamp }}

共分析 {{ test_count }} 个测试项,{{ total_points }} 个数据点,来自 {{ file_count }} 个文件

📁 处理的文件列表

{% for file_info in file_infos %}
{{ loop.index }}. {{ file_info.filename }}
路径: {{ file_info.path }}
数据行数: {{ file_info.rows }} | 测试项数: {{ file_info.tests }}
{% endfor %}
{% if comparison_plot_images and comparison_plot_images|length > 0 %}
🔍 多关键词对比视图
{% for comparison_plot in comparison_plot_images %}
{{ comparison_plot.title }}
{{ comparison_plot.title }}
{% endfor %}
{% endif %} {% for test in tests %}
📋 {{ test.name }}
{{ test.status_display }}
数据点数
{{ test.stats.count }}
平均值
{{ "%.4f"|format(test.stats.mean) }}
中位数
{{ "%.4f"|format(test.stats.median) }}
标准差
{{ "%.4f"|format(test.stats.std) }}
最小值
{{ "%.4f"|format(test.stats.min) }}
最大值
{{ "%.4f"|format(test.stats.max) }}
{% if test.limits.lower is not none or test.limits.upper is not none %}
{% if test.limits.lower is not none %}
下限值
{{ "%.4f"|format(test.limits.lower) }}
{% endif %} {% if test.limits.upper is not none %}
上限值
{{ "%.4f"|format(test.limits.upper) }}
{% endif %}
{% endif %}
📈 汇总视图 (所有SN)
{{ test.name }} 汇总散点图
{% if test.sn_plot_images %}
🔍 SN独立视图 ({{ test.sn_plot_images|length }}个SN)
{% for sn_plot in test.sn_plot_images %}
SN: {{ sn_plot.sn }}
{{ test.name }} - SN {{ sn_plot.sn }} 散点图
{% endfor %}
{% endif %}
{% endfor %}

📈 分析摘要

文件夹路径: {{ folder_path }}
分析时间: {{ analysis_time }}秒
关键词: {% for keyword in keywords %} {{ keyword }} {% endfor %}
测试项分布:
数据摘要:
报告生成于 {{ timestamp }} | 多文件测试报告分析系统
""" class MultiFileTestReportScatterPlotter: def __init__(self): self.folder_path: Optional[str] = None self.df: Optional[pd.DataFrame] = None self.output_dir: Optional[str] = None self.required_columns = ["Test Name New", "SN", "Measurement", "Test Time", "Lower Limit", "Upper Limit"] self.col_lower: Optional[str] = None self.col_upper: Optional[str] = None self.html_report_path: Optional[str] = None self.file_infos: List[Dict[str, Any]] = [] # 缓存处理过的数据 self._processed_data_cache: Dict[str, Any] = {} # 性能监控 self.performance_stats = { 'load_times': [], 'memory_usage': [], 'file_sizes': [] } def _print_stage(self, msg: str, color=Fore.CYAN) -> None: """统一的阶段信息输出""" print(f"\n{color}{'=' * 50}") print(f"📋 {msg}") print(f"{'=' * 50}{Style.RESET_ALL}") def _print_progress(self, current: int, total: int, prefix: str = "进度", color=Fore.YELLOW) -> None: """改进的进度条显示""" if total <= 0: return percent = (current / total) * 100 bar_len = 40 filled = int(bar_len * current / total) bar = "█" * filled + "░" * (bar_len - filled) sys.stdout.write(f"\r{color}{prefix}: [{bar}] {current}/{total} ({percent:.1f}%){Style.RESET_ALL}") sys.stdout.flush() if current == total: print(f"{Fore.GREEN} ✅ 完成{Style.RESET_ALL}") def _print_warning(self, msg: str) -> None: """警告信息输出""" print(f"{Fore.YELLOW}⚠️ {msg}{Style.RESET_ALL}") def _print_success(self, msg: str) -> None: """成功信息输出""" print(f"{Fore.GREEN}✅ {msg}{Style.RESET_ALL}") def _print_error(self, msg: str) -> None: """错误信息输出""" print(f"{Fore.RED}❌ {msg}{Style.RESET_ALL}") def _get_memory_usage(self) -> float: """获取当前内存使用量(GB)""" process = psutil.Process() return process.memory_info().rss / (1024 ** 3) def _check_memory_safe(self, file_size_mb: float) -> bool: """检查内存是否安全""" available_memory = psutil.virtual_memory().available / (1024 ** 3) estimated_need = file_size_mb * 5 / 1024 # 估算需要的内存(GB) return available_memory > estimated_need + 1 # 保留1GB安全空间 def _load_single_file_optimized(self, file_info: Dict[str, Any]) -> Optional[pd.DataFrame]: """优化单文件加载方法""" file_path = file_info['path'] filename = file_info['filename'] try: start_time = time.time() file_size_mb = os.path.getsize(file_path) / (1024 ** 2) # 内存安全检查 if not self._check_memory_safe(file_size_mb): self._print_warning(f"内存不足,跳过大文件: {filename} ({file_size_mb:.1f}MB)") return None # 选择合适的引擎 file_ext = file_path.lower() if file_ext.endswith('.xlsx'): engine = 'openpyxl' elif file_ext.endswith('.xls'): engine = 'xlrd' else: self._print_warning(f"不支持的文件格式: {filename}") return None # 快速获取工作表信息 try: excel_file = pd.ExcelFile(file_path, engine=engine) sheet_names = excel_file.sheet_names # 选择工作表 target_sheets = ["Merged All Tests", "All Tests", sheet_names[0] if sheet_names else None] selected_sheet = next((s for s in target_sheets if s and s in sheet_names), None) if not selected_sheet: self._print_warning(f"未找到目标工作表: {filename}") return None except Exception as e: self._print_warning(f"无法读取工作表信息 {filename}: {e}") return None # 优化读取参数 read_kwargs = { 'io': file_path, 'sheet_name': selected_sheet, 'engine': engine, 'dtype': 'object', 'na_filter': False, 'usecols': self.required_columns, # 只读取需要的列 } # 对于大文件,使用分块读取 if file_size_mb > 50: # 50MB以上使用分块读取 chunks = [] for chunk in pd.read_excel(**read_kwargs, chunksize=OPTIMIZATION_CONFIG['chunk_size']): chunks.append(chunk) if chunks: df = pd.concat(chunks, ignore_index=True) else: df = pd.DataFrame() else: df = pd.read_excel(**read_kwargs) if df.empty: self._print_warning(f"文件为空: {filename}") return None # 检查必要列 missing_columns = [col for col in self.required_columns if col not in df.columns] if missing_columns: self._print_warning(f"缺少必要列 {filename}: {missing_columns}") return None # 添加文件标识 df['_source_file'] = filename load_time = time.time() - start_time file_info.update({ 'load_time': round(load_time, 2), 'file_size_mb': round(file_size_mb, 2), 'engine': engine, 'rows': len(df) }) self.performance_stats['load_times'].append(load_time) self.performance_stats['file_sizes'].append(file_size_mb) self.performance_stats['memory_usage'].append(self._get_memory_usage()) self._print_success(f"加载完成: {filename} ({len(df)}行, {load_time:.2f}s)") return df except Exception as e: self._print_error(f"加载文件失败 {filename}: {e}") return None def _find_column_case_insensitive(self, candidates: List[str]) -> Optional[str]: """优化的大小写不敏感列查找""" if self.df is None: return None columns_lower = {col.lower().strip(): col for col in self.df.columns} for candidate in candidates: key = candidate.lower().strip() if key in columns_lower: return columns_lower[key] return None # 以下方法保持不变(为节省空间省略部分重复代码) def get_folder_path(self) -> None: """获取文件夹路径""" self._print_stage("输入文件夹路径") while True: print(f"{Fore.WHITE}请输入包含Excel文件的文件夹路径: ") folder_path = input("> ").strip() if not folder_path: continue path_obj = Path(folder_path) if path_obj.exists() and path_obj.is_dir(): self.folder_path = str(path_obj.resolve()) print(f"{Fore.GREEN}已选择文件夹: {self.folder_path}{Style.RESET_ALL}") break else: self._print_error(f"文件夹不存在: {folder_path},请重新输入") def find_excel_files(self) -> List[str]: """查找文件夹中的所有Excel文件""" self._print_stage("扫描Excel文件") excel_files = [] valid_extensions = ('.xlsx', '.xls') try: for file_path in Path(self.folder_path).rglob('*'): if file_path.suffix.lower() in valid_extensions and file_path.is_file(): excel_files.append(str(file_path.resolve())) # 按文件名排序 excel_files.sort() self._print_success(f"找到 {len(excel_files)} 个Excel文件") for i, file_path in enumerate(excel_files, 1): print(f" {i:2d}. {os.path.basename(file_path)}") return excel_files except Exception as e: self._print_error(f"扫描文件夹时发生错误: {e}") return [] def load_multiple_files_optimized(self, excel_files: List[str]) -> None: """优化多文件加载方法""" self._print_stage("并行加载Excel文件") start_time = time.time() # 准备文件信息 file_infos = [{'path': path, 'filename': os.path.basename(path)} for path in excel_files] all_dataframes = [] self.file_infos = [] if OPTIMIZATION_CONFIG['use_threading'] and len(excel_files) > 1: # 使用多线程并行加载 with ThreadPoolExecutor(max_workers=OPTIMIZATION_CONFIG['max_workers']) as executor: futures = {executor.submit(self._load_single_file_optimized, file_info): file_info for file_info in file_infos} completed = 0 for future in futures: try: df = future.result(timeout=300) # 5分钟超时 if df is not None: all_dataframes.append(df) self.file_infos.append(futures[future]) completed += 1 self._print_progress(completed, len(excel_files), "并行加载文件") except Exception as e: file_info = futures[future] self._print_error(f"加载失败 {file_info['filename']}: {e}") else: # 顺序加载 for i, file_info in enumerate(file_infos, 1): self._print_progress(i, len(excel_files), "加载文件") df = self._load_single_file_optimized(file_info) if df is not None: all_dataframes.append(df) self.file_infos.append(file_info) if not all_dataframes: raise ValueError("没有成功加载任何Excel文件") # 合并数据 self._print_stage("合并数据") merge_start = time.time() try: self.df = pd.concat(all_dataframes, ignore_index=True, sort=False) merge_time = time.time() - merge_start total_time = time.time() - start_time avg_load_time = np.mean(self.performance_stats['load_times']) if self.performance_stats['load_times'] else 0 self._print_success(f"合并完成: {len(self.df)}行, {len(all_dataframes)}个文件") self._print_success(f"加载耗时: {total_time:.2f}s (平均: {avg_load_time:.2f}s/文件)") self._print_success(f"合并耗时: {merge_time:.2f}s") # 显示性能统计 print(f"\n{Fore.CYAN}📊 性能统计:") print(f" 平均加载时间: {avg_load_time:.2f}s") print(f" 峰值内存使用: {max(self.performance_stats['memory_usage']):.2f}GB") print(f" 总文件大小: {sum(self.performance_stats['file_sizes']):.1f}MB{Style.RESET_ALL}") except Exception as e: self._print_error(f"合并数据失败: {e}") raise # 记录上下限列名 self.col_lower = self._find_column_case_insensitive([ "Lower Limit", "lower limit", "lower_limit", "ll", "lower" ]) self.col_upper = self._find_column_case_insensitive([ "Upper Limit", "upper limit", "upper_limit", "ul", "upper" ]) def get_keywords(self) -> Tuple[pd.DataFrame, List[str], List[str]]: """获取用户输入的关键词并筛选数据 - 修改为支持多个关键词""" self._print_stage("筛选关键词") while True: keyword_input = input("请输入一个或多个关键词(用逗号分隔,匹配 'Test Name New'): ").strip() if not keyword_input: print("❌ 关键词不能为空,请重新输入") continue # 分割关键词 keywords = [k.strip() for k in keyword_input.split(',') if k.strip()] if not keywords: print("❌ 没有有效的关键词,请重新输入") continue # 检查数据框是否为空 if self.df is None or self.df.empty: print("⚠️ 数据框为空,无法进行筛选") return pd.DataFrame(), keywords, [] # 检查列是否存在 if "Test Name New" not in self.df.columns: print("❌ 列 'Test Name New' 不存在于数据框中") print(f"可用列: {list(self.df.columns)}") return pd.DataFrame(), keywords, [] try: # 创建多个关键词的筛选条件 mask = pd.Series(False, index=self.df.index) for keyword in keywords: keyword_mask = self.df["Test Name New"].astype(str).str.contains(keyword, case=False, na=False) mask = mask | keyword_mask filtered_df = self.df.loc[mask].copy() if filtered_df.empty: # 提供友好的提示和建议 print(f"⚠️ 没有找到包含关键词 '{', '.join(keywords)}' 的测试项") # 显示部分可用的测试项作为参考 available_tests = self.df["Test Name New"].dropna().unique() if len(available_tests) > 0: print("📋 可用的测试项示例:") for test in available_tests[:5]: print(f" - {test}") if len(available_tests) > 5: print(f" ... 还有 {len(available_tests) - 5} 个测试项") # 提供重新输入或退出的选项 choice = input("请选择: 1-重新输入关键词 2-使用所有数据 3-退出当前操作: ") if choice == "1": continue elif choice == "2": filtered_df = self.df.copy() unique_tests = filtered_df["Test Name New"].unique().tolist() print(f"✅ 使用所有数据: {len(filtered_df)} 行,{len(unique_tests)} 个测试项") return filtered_df, keywords, unique_tests else: print("👋 退出筛选操作") return pd.DataFrame(), keywords, [] else: unique_tests = filtered_df["Test Name New"].unique().tolist() print(f"✅ 匹配到 {len(filtered_df)} 行数据,涉及 {len(unique_tests)} 个不同测试项") print(f" 使用的关键词: {', '.join(keywords)}") return filtered_df, keywords, unique_tests except Exception as e: print(f"❌ 筛选过程中发生错误: {e}") print("请检查数据格式或重新输入关键词") continue def create_output_dir(self, keyword) -> None: """创建输出目录""" self._print_stage("创建输出目录") if not self.folder_path: raise ValueError("文件夹路径未设置") timestamp = datetime.now().strftime("%Y%m%d_%H%M%S") self.output_dir = os.path.join(self.folder_path, f"scatter_report_out") safe_keyword = self._safe_filename(keyword) if keyword else "all_data" self.html_report_path = os.path.join(self.output_dir, f"{safe_keyword}_report_{timestamp}.html") os.makedirs(self.output_dir, exist_ok=True) print(f"输出目录: {self.output_dir}") @staticmethod def _safe_filename(name: str) -> str: """生成安全的文件名""" safe = "".join(c for c in str(name) if c.isalnum() or c in (" ", "_", "-")).strip() return safe or "Unknown_Test" def _extract_limits(self, df_one_test: pd.DataFrame) -> Tuple[ Optional[float], Optional[float], List[float], List[float]]: """提取某个测试项的上下限数值""" lower_plot = upper_plot = None lower_set = [] upper_set = [] if self.col_lower and self.col_lower in df_one_test.columns: lower_vals = self._clean_and_convert_series(df_one_test[self.col_lower], 'numeric').dropna().unique() lower_set = sorted(lower_vals.tolist()) if len(lower_vals) > 0 else [] if lower_set: lower_plot = min(lower_set) if self.col_upper and self.col_upper in df_one_test.columns: upper_vals = self._clean_and_convert_series(df_one_test[self.col_upper], 'numeric').dropna().unique() upper_set = sorted(upper_vals.tolist()) if len(upper_vals) > 0 else [] if upper_set: upper_plot = max(upper_set) return lower_plot, upper_plot, lower_set, upper_set @staticmethod def _clean_and_convert_series(series: pd.Series, target_type: str = 'numeric') -> pd.Series: """统一的系列清洗和转换方法 - 修复了 ast 方法名错误""" if series.empty: return series if target_type == 'numeric': # 数值转换优化 if pd.api.types.is_numeric_dtype(series): return series.astype(float) # 批量字符串处理 - 修复这里的问题 cleaned = series.astype(str).str.replace(r'[, ]', '', regex=True).str.strip() return pd.to_numeric(cleaned, errors='coerce') elif target_type == 'datetime': return MultiFileTestReportScatterPlotter._convert_to_datetime(series) return series @staticmethod def _convert_to_datetime(series: pd.Series) -> pd.Series: """优化的日期时间转换""" if pd.api.types.is_datetime64_any_dtype(series): return series # 预处理:转换为数值和字符串两种形式 numeric_series = pd.to_numeric(series, errors='coerce') string_series = series.astype(str).str.strip() result = pd.Series(pd.NaT, index=series.index, dtype='datetime64[ns]') # 数值时间戳处理 - 优化逻辑 masks = { 'ms': numeric_series >= 1e12, # 调整为更合理的阈值 's_ms': (numeric_series >= 1e9) & (numeric_series < 1e12), # 包含秒和毫秒的混合情况 'excel': (numeric_series > 20000) & (numeric_series < 60000) } for mask_type, mask in masks.items(): if mask.any(): if mask_type == 'ms': result.loc[mask] = pd.to_datetime(numeric_series.loc[mask], unit='ms') elif mask_type == 's_ms': # 对有小数部分的时间戳使用浮点数处理 timestamp_values = numeric_series.loc[mask] # 检查是否有小数部分 has_decimal = (timestamp_values % 1 != 0) # 对整数部分(秒级时间戳)处理 if (~has_decimal).any(): integer_mask = mask & (~has_decimal) result.loc[integer_mask] = pd.to_datetime( numeric_series.loc[integer_mask].astype('int64'), unit='s' ) # 对小数部分(可能是毫秒级)处理 if has_decimal.any(): decimal_mask = mask & has_decimal # 尝试毫秒单位转换 result.loc[decimal_mask] = pd.to_datetime( numeric_series.loc[decimal_mask] * 1000, unit='ms' ) elif mask_type == 'excel': origin = pd.Timestamp('1899-12-30') result.loc[mask] = origin + pd.to_timedelta(numeric_series.loc[mask], unit='D') # 字符串日期处理 remaining_mask = result.isna() if remaining_mask.any(): remaining_strings = string_series.loc[remaining_mask] # 特定格式优先处理 format_patterns = [ (r'^\d{4}-\d{2}-\d{2} \d{2}-\d{2}-\d{2}$', '%Y-%m-%d %H-%M-%S'), ] for pattern, date_format in format_patterns: format_mask = remaining_strings.str.match(pattern) if format_mask.any(): result.loc[remaining_mask[remaining_mask].index[format_mask]] = pd.to_datetime( remaining_strings.loc[format_mask], format=date_format, errors='coerce' ) # 通用解析 still_na_mask = result.isna() & remaining_mask if still_na_mask.any(): result.loc[still_na_mask] = pd.to_datetime( string_series.loc[still_na_mask], errors='coerce' ) return result def _preprocess_test_data(self, test_data: pd.DataFrame) -> pd.DataFrame: """数据预处理""" # 数值转换 test_data['Measurement_num'] = self._clean_and_convert_series( test_data['Measurement'], 'numeric' ) test_data['TestTime_dt'] = self._clean_and_convert_series( test_data['Test Time'], 'datetime' ) # 去除无效数据 valid_data = test_data.dropna(subset=['Measurement_num', 'TestTime_dt']) return valid_data.sort_values('TestTime_dt') def _preprocess_data(self, test_data: pd.DataFrame) -> pd.DataFrame: """数据预处理 - 简化版本用于对比图""" test_data = test_data.copy() # 数值转换 test_data['Measurement_num'] = self._clean_and_convert_series( test_data['Measurement'], 'numeric' ) test_data['TestTime_dt'] = self._clean_and_convert_series( test_data['Test Time'], 'datetime' ) # 去除无效数据 valid_data = test_data.dropna(subset=['Measurement_num', 'TestTime_dt']) return valid_data def _calculate_statistics(self, y_data: pd.Series) -> Dict[str, float]: """计算统计信息""" stats = { 'count': len(y_data), 'mean': y_data.mean(), 'median': y_data.median(), 'min': y_data.min(), 'max': y_data.max(), 'std': y_data.std(), 'q1': y_data.quantile(0.25), 'q3': y_data.quantile(0.75) } return stats def _plot_to_base64(self, fig) -> str: """将图表转换为base64编码""" buf = BytesIO() fig.savefig(buf, format='png', dpi=150, bbox_inches='tight') buf.seek(0) img_str = base64.b64encode(buf.read()).decode('utf-8') plt.close(fig) return img_str def _create_summary_plot(self, test_data: pd.DataFrame, test_name: str, lower_plot: Optional[float], upper_plot: Optional[float]) -> str: """创建汇总图(所有SN在一个图中)""" fig, ax = plt.subplots(figsize=(12, 8)) # 分组绘制 groups = list(test_data.groupby("SN")) if "SN" in test_data.columns else [("Unknown_SN", test_data)] for sn, group in groups: ax.scatter(group['TestTime_dt'], group['Measurement_num'], label=str(sn), alpha=0.7, s=25) # 计算统计信息 y_data = test_data['Measurement_num'] stats = self._calculate_statistics(y_data) # 绘制限值线和统计线 x_min, x_max = test_data['TestTime_dt'].min(), test_data['TestTime_dt'].max() if lower_plot is not None: ax.axhline(y=lower_plot, color='green', linestyle='--', linewidth=1.2, label="Lower Limit") if upper_plot is not None: ax.axhline(y=upper_plot, color='red', linestyle='--', linewidth=1.2, label="Upper Limit") # 添加统计线 ax.hlines(y=stats['mean'], xmin=x_min, xmax=x_max, colors='orange', linestyles='-', linewidth=1.5, alpha=0.7, label='Mean') ax.hlines(y=stats['median'], xmin=x_min, xmax=x_max, colors='purple', linestyles='-.', linewidth=1.5, alpha=0.7, label='Median') # 设置图形属性 ax.set_title(f"汇总图 - {test_name}") ax.set_xlabel("Test Time") ax.set_ylabel("Measurement Value") ax.grid(True, alpha=0.3) ax.tick_params(axis='x', rotation=45) ax.legend(bbox_to_anchor=(1.05, 1), loc='upper left') return self._plot_to_base64(fig) def _create_sn_plots(self, test_data: pd.DataFrame, test_name: str, lower_plot: Optional[float], upper_plot: Optional[float]) -> List[Dict[str, str]]: """为每个SN创建独立图表""" sn_plots = [] if "SN" not in test_data.columns: return sn_plots sn_groups = test_data.groupby("SN") for sn, group in sn_groups: if group.empty: continue fig, ax = plt.subplots(figsize=(10, 6)) # 绘制当前SN的数据点 ax.scatter(group['TestTime_dt'], group['Measurement_num'], color='blue', alpha=0.7, s=30, label=f"SN: {sn}") # 计算当前SN的统计信息 y_data = group['Measurement_num'] stats = self._calculate_statistics(y_data) # 绘制限值线 x_min, x_max = group['TestTime_dt'].min(), group['TestTime_dt'].max() if lower_plot is not None: ax.axhline(y=lower_plot, color='green', linestyle='--', linewidth=1.2, label="Lower Limit") if upper_plot is not None: ax.axhline(y=upper_plot, color='red', linestyle='--', linewidth=1.2, label="Upper Limit") # 添加统计线 ax.hlines(y=stats['mean'], xmin=x_min, xmax=x_max, colors='orange', linestyles='-', linewidth=1.5, alpha=0.7, label='Mean') ax.hlines(y=stats['median'], xmin=x_min, xmax=x_max, colors='purple', linestyles='-.', linewidth=1.5, alpha=0.7, label='Median') # 设置图形属性 ax.set_title(f"SN独立图 - {test_name} (SN: {sn})") ax.set_xlabel("Test Time") ax.set_ylabel("Measurement Value") ax.grid(True, alpha=0.3) ax.tick_params(axis='x', rotation=45) ax.legend() # 转换为base64 plot_image = self._plot_to_base64(fig) sn_plots.append({"sn": str(sn), "image": plot_image}) return sn_plots def _determine_test_status(self, stats: Dict[str, float], lower_limit: Optional[float], upper_limit: Optional[float]) -> Dict[str, Any]: """确定测试状态""" status = "success" status_display = "正常" if lower_limit is not None and upper_limit is not None: # 检查是否超出限值 if stats['min'] < lower_limit or stats['max'] > upper_limit: status = "danger" status_display = "异常" elif (stats['mean'] < lower_limit * 1.1 or stats['mean'] > upper_limit * 0.9 or stats['std'] > (upper_limit - lower_limit) * 0.2): status = "warning" status_display = "警告" return {"status": status, "status_display": status_display} def _create_comparison_plots(self, filtered_df: pd.DataFrame, keywords: List[str]) -> List[Dict[str, str]]: """创建多关键词对比图 - 优化版本:时间序列改为散点图""" comparison_plots = [] if len(keywords) <= 1: return comparison_plots # 单一关键词不需要对比图 try: # 确保有足够的数据 if filtered_df.empty: return comparison_plots # 1. 散点图对比(原时间序列图改为散点图) fig1, ax1 = plt.subplots(figsize=(14, 8)) has_data = False # 使用更鲜艳的颜色 colors = plt.cm.Set3(np.linspace(0, 1, len(keywords))) markers = ['o', 's', '^', 'D', 'v', '<', '>', 'p', '*', 'h'] # 多种标记形状 # 获取全局时间范围用于统一x轴 global_min_time = None global_max_time = None # 先收集所有数据的时间范围 for keyword in keywords: keyword_mask = filtered_df["Test Name New"].astype(str).str.contains( re.escape(keyword), case=False, na=False ) keyword_data = filtered_df.loc[keyword_mask].copy() if not keyword_data.empty: keyword_data = self._preprocess_test_data(keyword_data) if not keyword_data.empty: time_min = keyword_data['TestTime_dt'].min() time_max = keyword_data['TestTime_dt'].max() if global_min_time is None or time_min < global_min_time: global_min_time = time_min if global_max_time is None or time_max > global_max_time: global_max_time = time_max for i, keyword in enumerate(keywords): keyword_mask = filtered_df["Test Name New"].astype(str).str.contains( re.escape(keyword), case=False, na=False ) keyword_data = filtered_df.loc[keyword_mask].copy() if not keyword_data.empty: keyword_data = self._preprocess_test_data(keyword_data) if not keyword_data.empty: # 对散点图数据添加少量随机抖动,避免完全重叠 if len(keyword_data) > 1: # 为相同时间点的数据添加微小的时间偏移,避免重叠 time_jitter = pd.Timedelta(minutes=1) # 1分钟抖动 jitter_range = np.random.uniform(-0.5, 0.5, len(keyword_data)) * time_jitter keyword_data['TestTime_dt_jittered'] = keyword_data['TestTime_dt'] + jitter_range x_values = keyword_data['TestTime_dt_jittered'] else: x_values = keyword_data['TestTime_dt'] y_values = keyword_data['Measurement_num'] if len(x_values) > 0: # 使用散点图,设置不同的标记和透明度 scatter = ax1.scatter(x_values, y_values, label=f"{keyword} (n={len(keyword_data)})", color=colors[i], marker=markers[i % len(markers)], s=40, # 点的大小 alpha=0.7, # 透明度 edgecolors='white', # 边缘颜色 linewidth=0.5) # 边缘线宽 has_data = True # 为每个关键词添加趋势线(可选) if len(keyword_data) >= 3: try: # 按时间排序 sorted_data = keyword_data.sort_values('TestTime_dt') # 计算简单线性趋势 x_numeric = pd.to_numeric(sorted_data['TestTime_dt']) y_trend = sorted_data['Measurement_num'] if len(x_numeric) >= 2: # 使用numpy的polyfit计算趋势线 z = np.polyfit(x_numeric, y_trend, 1) p = np.poly1d(z) ax1.plot(sorted_data['TestTime_dt'], p(x_numeric), color=colors[i], linestyle='--', alpha=0.5, linewidth=1, label=f"{keyword} 趋势线") except: pass # 趋势线计算失败时忽略 if has_data: # 设置统一的x轴范围 if global_min_time and global_max_time: # 添加一些边距 time_range = global_max_time - global_min_time margin = time_range * 0.05 ax1.set_xlim(global_min_time - margin, global_max_time + margin) ax1.set_title("多关键词散点图对比", fontsize=14, fontweight='bold') ax1.set_xlabel("测试时间", fontsize=12) ax1.set_ylabel("测量值", fontsize=12) ax1.grid(True, alpha=0.3) ax1.tick_params(axis='x', rotation=45) # 优化图例显示 ax1.legend(bbox_to_anchor=(1.05, 1), loc='upper left', fontsize=10, framealpha=0.9) plt.tight_layout() comparison_plots.append({ "title": "散点图对比", "image": self._plot_to_base64(fig1) }) # 2. 箱线图对比(简化版) plot_data = [] labels = [] for keyword in keywords: keyword_mask = filtered_df["Test Name New"].astype(str).str.contains( re.escape(keyword), case=False, na=False ) keyword_data = filtered_df.loc[keyword_mask].copy() if not keyword_data.empty: keyword_data = self._preprocess_test_data(keyword_data) if not keyword_data.empty and len(keyword_data) >= 3: # 至少3个数据点 plot_data.append(keyword_data['Measurement_num'].values) labels.append(f"{keyword}\n(n={len(keyword_data)})") if len(plot_data) >= 2: # 至少两个关键词有数据 fig2, ax2 = plt.subplots(figsize=(10, 6)) bp = ax2.boxplot(plot_data, tick_labels=labels, patch_artist=True) colors = plt.cm.Set3(np.linspace(0, 1, len(plot_data))) for i, box in enumerate(bp['boxes']): box.set(facecolor=colors[i], alpha=0.7) ax2.set_title("多关键词箱线图对比", fontsize=14, fontweight='bold') ax2.set_ylabel("测量值", fontsize=12) ax2.grid(True, alpha=0.3) plt.tight_layout() comparison_plots.append({ "title": "箱线图对比", "image": self._plot_to_base64(fig2) }) # 3. 分布直方图对比 fig3, ax3 = plt.subplots(figsize=(12, 6)) has_hist_data = False colors = plt.cm.Set3(np.linspace(0, 1, len(keywords))) for i, keyword in enumerate(keywords): keyword_mask = filtered_df["Test Name New"].astype(str).str.contains( re.escape(keyword), case=False, na=False ) keyword_data = filtered_df.loc[keyword_mask].copy() if not keyword_data.empty: keyword_data = self._preprocess_test_data(keyword_data) if not keyword_data.empty: # 动态调整直方图bins数量 n_bins = min(20, max(5, len(keyword_data) // 10)) ax3.hist(keyword_data['Measurement_num'].values, bins=n_bins, alpha=0.6, label=f"{keyword} (n={len(keyword_data)})", color=colors[i], density=True) # 使用密度而不是频次 has_hist_data = True if has_hist_data: ax3.set_title("多关键词分布直方图对比", fontsize=14, fontweight='bold') ax3.set_xlabel("测量值", fontsize=12) ax3.set_ylabel("密度", fontsize=12) ax3.legend() ax3.grid(True, alpha=0.3) plt.tight_layout() comparison_plots.append({ "title": "分布直方图对比", "image": self._plot_to_base64(fig3) }) except Exception as e: self._print_warning(f"创建对比图时出错: {e}") import traceback traceback.print_exc() return comparison_plots def generate_html_report(self, filtered_df: pd.DataFrame, keywords: List[str], unique_tests: List[str]) -> None: """生成HTML报告 - 修复对比图显示问题""" self._print_stage("生成HTML报告") start_time = time.time() test_results = [] total_points = 0 status_counts = {"success": 0, "warning": 0, "danger": 0} # 生成多关键词对比图 print(f"🔍 调试: 开始生成对比图,关键词数量: {len(keywords)}") comparison_plot_images = self._create_comparison_plots(filtered_df, keywords) print(f"🔍 调试: 对比图生成完成,数量: {len(comparison_plot_images)}") # 调试输出对比图信息 for i, plot in enumerate(comparison_plot_images): print(f" - 对比图{i + 1}: {plot['title']}, 图像大小: {len(plot['image'])} 字符") # 生成各测试项的详细图表 for i, test_name in enumerate(unique_tests, 1): self._print_progress(i, len(unique_tests), "生成测试报告") # 获取测试数据 test_data = filtered_df[filtered_df["Test Name New"] == test_name].copy() test_data = self._preprocess_test_data(test_data) if test_data.empty: continue # 提取限值信息 lower_plot, upper_plot, _, _ = self._extract_limits(test_data) # 计算统计信息 y_data = test_data['Measurement_num'] stats = self._calculate_statistics(y_data) total_points += stats['count'] # 生成汇总图表 summary_plot_image = self._create_summary_plot(test_data, test_name, lower_plot, upper_plot) # 生成SN独立图表 sn_plot_images = self._create_sn_plots(test_data, test_name, lower_plot, upper_plot) # 确定测试状态 status_info = self._determine_test_status(stats, lower_plot, upper_plot) status_counts[status_info["status"]] += 1 # 添加到结果列表 test_results.append({ "name": test_name, "stats": stats, "limits": {"lower": lower_plot, "upper": upper_plot}, "summary_plot_image": summary_plot_image, "sn_plot_images": sn_plot_images, "status": status_info["status"], "status_display": status_info["status_display"] }) # 渲染HTML模板前再次验证数据 print(f"🔍 调试: 传递给模板的对比图数量: {len(comparison_plot_images)}") # 渲染HTML模板 template = Template(HTML_TEMPLATE) html_content = template.render( keywords=keywords if keywords else ["所有数据"], keywords_display=", ".join(keywords) if keywords else "所有数据", timestamp=datetime.now().strftime("%Y-%m-%d %H:%M:%S"), test_count=len(test_results), total_points=total_points, tests=test_results, folder_path=self.folder_path, analysis_time=round(time.time() - start_time, 2), status_counts={"normal": status_counts["success"], "warning": status_counts["warning"], "abnormal": status_counts["danger"]}, file_count=len(self.file_infos), file_infos=self.file_infos, total_rows=len(self.df) if self.df is not None else 0, comparison_plot_images=comparison_plot_images # 确保传递 ) # 调试:检查生成的HTML内容 if comparison_plot_images: if "comparison_plot_images" in html_content or "时间序列对比图" in html_content: print("✅ 对比图已成功嵌入HTML") else: print("❌ 对比图未正确嵌入HTML") # 保存HTML文件 with open(self.html_report_path, 'w', encoding='utf-8') as f: f.write(html_content) self._print_success(f"HTML报告已生成: {self.html_report_path}") self._print_success( f"共处理 {len(self.file_infos)} 个文件,{len(test_results)} 个测试项,{total_points} 个数据点") if len(keywords) > 1: self._print_success(f"已生成 {len(comparison_plot_images)} 个对比图表") def run(self) -> None: """运行主程序""" try: self.get_folder_path() excel_files = self.find_excel_files() if not excel_files: self._print_error("没有找到可用的Excel文件") return # 使用优化后的加载方法 self.load_multiple_files_optimized(excel_files) while True: filtered_df, keywords, unique_tests = self.get_keywords() # 修改方法调用 if filtered_df.empty: self._print_warning("没有数据可处理,退出程序") break safe_keyword_text = "_".join([self._safe_filename(k) for k in keywords]) if keywords else "all_data" self.create_output_dir(safe_keyword_text) self.generate_html_report(filtered_df, keywords, unique_tests) # 修改参数 self._print_success("分析完成!") print(f"📊 报告文件: {self.html_report_path}") print(f"📁 输出目录: {self.output_dir}") if len(keywords) > 1: print(f"🔍 对比关键词: {', '.join(keywords)}") # 询问是否继续分析其他关键词 continue_choice = input("\n是否继续分析其他关键词?(y/n): ").strip().lower() if continue_choice not in ['y', 'yes', '是']: break except KeyboardInterrupt: self._print_warning("用户中断程序") except Exception as e: self._print_error(f"发生错误: {type(e).__name__}: {str(e)}") import traceback traceback.print_exc() if __name__ == "__main__": plotter = MultiFileTestReportScatterPlotter() plotter.run()