| | import json |
| | import datetime |
| | import pandas as pd |
| | from difflib import SequenceMatcher |
| |
|
| | def string_similarity(str1, str2): |
| | |
| | if (str1 is None or str1 == "") and (str2 is not None and str2 != ""): |
| | return 0.0 |
| | if (str2 is None or str2 == "") and (str1 is not None and str1 != ""): |
| | return 0.0 |
| |
|
| | |
| | if (str1 or "") == (str2 or ""): |
| | return 1.0 |
| |
|
| |
|
| | |
| | s1_lower = str1.lower() |
| | s2_lower = str2.lower() |
| |
|
| | |
| | if s1_lower == s2_lower: |
| | return 1.0 |
| |
|
| | |
| | matcher = SequenceMatcher(None, s1_lower, s2_lower) |
| | similarity = matcher.ratio() |
| |
|
| | return similarity |
| |
|
| | def main(file_llm = '', file_bench = '', key_list = []): |
| | all_data = {} |
| | with open(file_llm, 'r', encoding='utf-8') as f: |
| | for line in f: |
| | j = json.loads(line).get('llm_response_dict') |
| | |
| | all_data[j['sha256']] = {} |
| | all_data[j['sha256']]['llm_response_dict'] = j |
| |
|
| | with open(file_bench, 'r', encoding='utf-8') as f: |
| | for line in f: |
| | j = json.loads(line) |
| | if j['sha256'] not in all_data: |
| | all_data[j['sha256']] = {} |
| | all_data[j['sha256']]['benchmark_dict'] = j |
| |
|
| | sha256_to_remove = [] |
| | for sha256, value in all_data.items(): |
| | |
| | if 'llm_response_dict' not in value or 'benchmark_dict' not in value: |
| | sha256_to_remove.append(sha256) |
| | for sha256 in sha256_to_remove: |
| | all_data.pop(sha256) |
| |
|
| | for sha256, value in all_data.items(): |
| | all_data[sha256]['similarity'] = {} |
| | for key in key_list: |
| | |
| | all_data[sha256]['similarity'][key] = string_similarity(all_data[sha256]['llm_response_dict'].get(key), all_data[sha256]['benchmark_dict'][key]) |
| |
|
| | |
| | key_accuracy_tmp = {key: 0 for key in key_list} |
| | for sha256, value in all_data.items(): |
| | for key in key_list: |
| | key_accuracy_tmp[key] += value['similarity'][key] |
| |
|
| | |
| | key_accuracy = {k: v / len(all_data) for k,v in key_accuracy_tmp.items()} |
| | |
| | accuracy = sum(list(key_accuracy.values())) / len(list(key_accuracy.values())) |
| | return accuracy, key_accuracy, all_data |
| |
|
| | def write_similarity_data_to_excel(key_list, data_dict, output_file="similarity_analysis.xlsx"): |
| | """ |
| | 将相似度分析数据写入Excel文件 |
| | |
| | Args: |
| | data_dict: 包含相似度分析数据的字典 |
| | output_file: 输出Excel文件名 |
| | """ |
| |
|
| | |
| | rows = [] |
| |
|
| | for sha256, data in data_dict.items(): |
| | row = { |
| | 'sha256': sha256 |
| | } |
| |
|
| | for field in key_list: |
| | |
| | row[f'llm_{field}'] = data['llm_response_dict'].get(field) |
| | |
| | row[f'benchmark_{field}'] = data['benchmark_dict'].get(field) |
| | |
| | row[f'similarity_{field}'] = data['similarity'].get(field) |
| | rows.append(row) |
| |
|
| | |
| | df = pd.DataFrame(rows) |
| |
|
| | |
| | column_order = ['sha256'] |
| | for field in key_list: |
| | column_order.extend([f'llm_{field}', f'benchmark_{field}', f'similarity_{field}']) |
| |
|
| | |
| | df = df[column_order] |
| |
|
| | |
| | with pd.ExcelWriter(output_file, engine='openpyxl') as writer: |
| | df.to_excel(writer, sheet_name='相似度分析', index=False) |
| |
|
| | |
| | |
| | |
| |
|
| |
|
| | print(f"数据已成功写入 {output_file}") |
| | print(f"总共处理了 {len(rows)} 条记录") |
| |
|
| | return df |
| |
|
| | if __name__ == '__main__': |
| | file_llm = 'data/llm-label_textbook.jsonl' |
| | file_bench = 'data/benchmark_textbook.jsonl' |
| | |
| | key_list = ['isbn', 'title', 'author', 'abstract', 'category', 'pub_time', 'publisher'] |
| |
|
| | accuracy, key_accuracy, detail_data = main(file_llm, file_bench, key_list) |
| | |
| | timestamp = datetime.datetime.now().strftime("%Y%m%d_%H%M%S") |
| | output_filename = f"similarity_analysis_{timestamp}.xlsx" |
| | write_similarity_data_to_excel(key_list, detail_data, output_filename) |
| |
|
| | print(key_accuracy) |
| | print(accuracy) |
| |
|
| |
|
| |
|
| |
|
| |
|