Skip to content

YaoJunHao init#7

Open
lfkg wants to merge 118 commits intomainfrom
yjh-58235-eval-dev
Open

YaoJunHao init#7
lfkg wants to merge 118 commits intomainfrom
yjh-58235-eval-dev

Conversation

@lfkg
Copy link

@lfkg lfkg commented Sep 11, 2025

No description provided.

chanzhi82020 and others added 30 commits August 25, 2025 16:57
…and benchmark applications built on the FastGPT platform. (labring#5476)

- Adds a lightweight evaluation framework for app-level tracking and benchmarking.
- Changes: 28 files, +1455 additions, -66 deletions.
- Branch: add-evaluations -> main.
- PR: chanzhi82020#1

Applications built on FastGPT need repeatable, comparable benchmarks to measure regressions, track improvements, and validate releases. This initial implementation provides the primitives to define evaluation scenarios, run them against app endpoints or model components, and persist results for later analysis.

I updated the PR description to emphasize that the evaluation system is targeted at FastGPT-built apps and expanded the explanation of the core pieces so reviewers understand the scope and intended use. The new description outlines the feature intent, core components, and how results are captured and aggregated for benchmarking.

- Evaluation definitions
  - Define evaluation tasks that reference an app (app id, version, endpoint), test datasets or input cases, expected outputs (when applicable), and run configuration (parallelism, timeouts).
  - Support for custom metric plugins so teams can add domain-specific measures.

- Runner / Executor
  - Executes evaluation cases against app endpoints or internal model interfaces.
  - Captures raw responses, response times, status codes, and any runtime errors.
  - Computes per-case metrics (e.g., correctness, latency) immediately after each case run.

- Metrics & Aggregation
  - Built-in metrics: accuracy/success rate, latency (p50/p90/p99), throughput, error rate.
  - Aggregation produces per-run summaries and per-app historical summaries for trend analysis.
  - Allows combining metrics into composite scores for high-level benchmarking.

- Persistence & Logging
  - Stores run results, input/output pairs (when needed), timestamps, environment info, and app/version metadata so runs are reproducible and auditable.
  - Logs are retained to facilitate debugging and root-cause analysis of regressions.

- Reporting & Comparison
  - Produces aggregated reports suitable for CI gating, release notes, or dashboards.
  - Supports comparing multiple app versions or deployments side-by-side.

- Extensibility & Integration
  - Designed to plug into CI (automated runs on PRs or releases), dashboards, and downstream analysis tools.
  - Easy to add new metrics, evaluators, or dataset connectors.

By centering the evaluation system on FastGPT apps, teams can benchmark full application behavior (not only raw model outputs), correlate metrics with deployment configurations, and make informed release decisions.

- Expand built-in metric suite (e.g., F1, BLEU/ROUGE where applicable), add dataset connectors, and provide example evaluation scenarios for sample apps.
- Integrate with CI pipelines and add basic dashboarding for trend visualization.

Related Issue: N/A

Co-authored-by: Archer <545436317@qq.com>
…eval-dataset-dev' 到 'eval-dev')

feat: Add comprehensive evaluation dataset management system

Summary

This PR introduces a comprehensive evaluation dataset management system that enables users to create, manage, and process evaluation datasets with advanced features including smart generation and quality assessment.

Key Features Added:

Dataset Collection Management: Complete CRUD operations for evaluation dataset collections with team/user isolation

Dataset Data Management: Complete CRUD operations for evaluation dataset data with team/user isolation

Smart Data Generation: AI-powered synthesis of evaluation data using background job processing

Quality Assessment: Batch quality evaluation jobs with configurable assessment criteria

Task Management: Queue-based processing system for long-running dataset operations

Technical Implementation:

Database Schemas: New MongoDB schemas for dataset collections and data entries with proper indexing

API Endpoints: RESTful APIs for all dataset operations with proper error handling and validation

Background Processing: BullMQ integration for handling intensive data processing tasks

Type Safety: Comprehensive TypeScript definitions and enums for improved data integrity

Team Integration: Full support for team-based access control and data isolation

New API Endpoints:

  Dataset Collection Management

  - POST /api/core/evaluation/dataset/collection/create - Create evaluation dataset collection
  - PUT /api/core/evaluation/dataset/collection/update - Update dataset collection
  - POST /api/core/evaluation/dataset/collection/list - List dataset collections with pagination
  - POST /api/core/evaluation/dataset/collection/failedTasks - Get failed processing tasks
  - POST /api/core/evaluation/dataset/collection/retryTask - Retry failed processing task
  - POST /api/core/evaluation/dataset/collection/deleteTask - Delete processing task

  Dataset Data Management

  - POST /api/core/evaluation/dataset/data/create - Create individual dataset data entry
  - PUT /api/core/evaluation/dataset/data/update - Update dataset data entry
  - POST /api/core/evaluation/dataset/data/list - List dataset data with pagination
  - DELETE /api/core/evaluation/dataset/data/delete - Delete dataset data entry

  Data Import & Processing

  - POST /api/common/file/upload - Upload CSV files for dataset import
  - POST /api/core/evaluation/dataset/data/fileId - Import dataset data from uploaded file

  Quality Assessment

  - POST /api/core/evaluation/dataset/data/qualityAssessment - Single data entry quality assessment
  - POST /api/core/evaluation/dataset/collection/qualityAssessmentBatch - Batch quality assessment for entire collection

  Smart Generation

  - POST /api/core/evaluation/dataset/data/smartGenerate - AI-powered smart generation of evaluation data from dataset

TODO
- [x] gridfs 文件上传需定义新的 bucket 专门提供给 evaluation 使用
- [ ] 认证和授权完善
- [ ] limit配额检查
- [ ] 计费统计
- [ ] 审计日志
- [ ] 异常错误码
- [ ] 调用 Diting 服务联调,包括质量评测和数据合成
- [x] 接口单元测试和集成测试
- [ ] 完善3个队列的并发配置确认和提供配置

设计文档: https://xcnxw5z29dc5.feishu.cn/base/BPIRbF5bsakhqIstp9ecp7K7ntc?table=ldx3tltRDjoWp0du

运行单元测试: pnpm exec vitest run test/cases/pages/api/core/evaluation/dataset/

@82020 @31202 @94619

查看合并请求 AI-PaaS/FastGPT!22
- Add evaluation dataset management page and detail page
- Add evaluation dimension management page
- Refactor evaluation homepage to tab layout, supporting task/dataset/dimension switching
- Add basic structure for evaluation task detail page
- Update navigation bar route configuration to support new page routes
…tures ('14864/evaluation-container' 到 'eval-dev')

feat: Add evaluation task, dataset and dimension management features

- Add evaluation dataset management page and detail page
- Add evaluation dimension management page
- Refactor evaluation homepage to tab layout, supporting task/dataset/dimension switching
- Add basic structure for evaluation task detail page
- Update navigation bar route configuration to support new page routes

查看合并请求 AI-PaaS/FastGPT!26
…ectors and form validation

- Merge multilingual files for evaluation dimensions and datasets
- Add evaluation dimension creation and editing pages
- Implement evaluation dimension form validation and submission logic
- Add evaluation dimension trial run functionality
- Optimize resource selector component, support hiding root directory and avatar display control
- Add reference template component, provide standard evaluation templates
- Implement answer input component, support collapse and automatic height adjustment
- Add application selector component, support displaying all application options
…translations' 到 'eval-dev')

[feat] updated some English translations

查看合并请求 AI-PaaS/FastGPT!32
…ze selectors and form validation ('14864/add-dimension' 到 'eval-dev')

feat: Add evaluation dimension management functionality, optimize selectors and form validation

- Merge multilingual files for evaluation dimensions and datasets
- Add evaluation dimension creation and editing pages
- Implement evaluation dimension form validation and submission logic
- Add evaluation dimension trial run functionality
- Optimize resource selector component, support hiding root directory and avatar display control
- Add reference template component, provide standard evaluation templates
- Implement answer input component, support collapse and automatic height adjustment
- Add application selector component, support displaying all application options

查看合并请求 AI-PaaS/FastGPT!30
…l-dev')

[feat]add evaluation plugin: diting

New Addition: Diting - Evaluation and Data Generation

Diting Core: Evaluation and Data Generation Engine
- Implemented the evaluation engine, supporting multiple evaluation metrics for application performance assessment.
- Integrated the data generation engine to meet the needs for synthetic datasets and application evaluations.

Diting Server: API Server for Evaluation and Data Generation
Added New API Endpoints
- POST /api/v1/evaluations/runs
  API interface for application evaluations.

- POST /api/v1/dataset-synthesis/runs
  API interface for data synthesis.
@82020 @64078 @10037

查看合并请求 AI-PaaS/FastGPT!25
…unctionality

- Add file import component with drag-and-drop upload and template download support
- Implement intelligent dataset generation modal with knowledge base selection and generation parameter configuration
- Add evaluation task creation modal with evaluation dimension and parameter configuration support
- Improve multi-language translation with new interface text additions
- Optimize file upload logic with progress display and error handling support
- Add evaluation dimension management component with dimension model selection and configuration support
- Implement evaluation parameter configuration modal with threshold and weight settings support
…tion functionality ('14864/evaluation-components' 到 'eval-dev')

feat: Add evaluation dataset file import and intelligent generation functionality

- Add file import component with drag-and-drop upload and template download support
- Implement intelligent dataset generation modal with knowledge base selection and generation parameter configuration
- Add evaluation task creation modal with evaluation dimension and parameter configuration support
- Improve multi-language translation with new interface text additions
- Optimize file upload logic with progress display and error handling support
- Add evaluation dimension management component with dimension model selection and configuration support
- Implement evaluation parameter configuration modal with threshold and weight settings support

查看合并请求 AI-PaaS/FastGPT!37
…tor-evaluation-backend' 到 'eval-dev')

feat: implement comprehensive backend evaluation system

Enhanced Evaluation System - Complete Architecture Refactor

This PR introduces a comprehensive evaluation framework designed specifically for tracking and benchmarking applications built on the FastGPT platform.

📋 Key Features Added

1. Multi-Component Architecture
- Evaluation Datasets: Structured data management with CSV/JSON support
- Evaluation Targets: Configurable workflow-based evaluation targets
- Evaluation Metrics: AI model-based evaluation with custom prompts
- Evaluation Tasks: Orchestrated evaluation execution with item tracking

2. Comprehensive Type System
- Enhanced API Types: 160+ new API interfaces for full CRUD operations
- Complex Schema Types: Dataset, Target, Metric, and Task type definitions
- Display Types: Optimized types for UI presentation
- Validation Types: Import/export and validation result handling

3. Advanced Queue Management
- Specialized Queues: evaluation_task and evaluation_item queues
- Parallel Processing: Concurrent evaluation item processing
- Error Handling: Comprehensive error status tracking
- Retry Logic: Built-in retry mechanisms for failed evaluations

5. Enhanced Status Management
- Added Error State: New error status for failed evaluations
- Status Tracking: Comprehensive status progression
- Progress Monitoring: Real-time evaluation progress

6. Chat Integration
- Evaluation Source: New chat source type for evaluation contexts
- Logging Support: Dedicated evaluation chat logging

🛠 Technical Improvements

Service Layer Enhancements
- Resource Validation: Unified resource access validation
- Permission Management: Team-based resource permissions
- Pagination Support: Consistent pagination across all endpoints
- Error Handling: Standardized error responses

Database Schema Updates
- Complex Schemas: Multi-level nested schema definitions
- Indexing Strategy: Optimized database indexes
- Relationship Management: Proper MongoDB relationships

Queue System Improvements
- Worker Management: Enhanced worker error handling
- Job Orchestration: Sophisticated job dependency management
- Background Processing: Efficient background task execution

📊 Architecture Benefits

1. Scalability: Modular design supports large-scale evaluations
2. Flexibility: Configurable evaluation targets and metrics
3. Reliability: Robust error handling and retry mechanisms
4. Maintainability: Clean separation of concerns
5. Extensibility: Plugin-like architecture for custom metrics

🏗 Architecture Overview

flowchart TD
    A[Dataset Management] --> B[Evaluation Task]
    C[Metric Configuration] --> B
    D[Target Configuration] --> B
    B --> E[Queue System]
    E --> F[Parallel Processing]
    F --> G[Result Aggregation]
    G --> H[Status Updates]

📋 Evaluation Task Execution Flow

sequenceDiagram
    participant User
    participant API
    participant Queue
    participant Worker
    participant Database

    User->>API: Create Evaluation Task
    API->>Database: Validate Components
    API->>Queue: Submit Task
    Queue->>Worker: Process Items
    Worker->>Database: Store Results
    Worker->>API: Update Status
    API->>User: Return Results

This represents a major architectural advancement from a basic evaluation system to a comprehensive, enterprise-grade evaluation platform with advanced features for dataset management, flexible metrics, and robust processing capabilities.

查看合并请求 AI-PaaS/FastGPT!38
… components

- Add EditDataModal component for editing evaluation questions and answers
- Implement ModifyEvaluationModal for modifying evaluation results
- Add evaluation status constants and mapping
- Support multiple evaluation statuses (HighQuality, NeedsImprovement, Abnormal)
- Add i18n translations for new evaluation features
- Include save, cancel and save+next functionality in edit modal
- Support manual evaluation result modification with reasons
- Add evaluation status badges and feedback display
…gement components ('76887/0904/eidt-modal' 到 'eval-dev')

feat(evaluation): add evaluation result editing and status management components

- Add EditDataModal component for editing evaluation questions and answers
- Implement ModifyEvaluationModal for modifying evaluation results
- Add evaluation status constants and mapping
- Support multiple evaluation statuses (HighQuality, NeedsImprovement, Abnormal)
- Add i18n translations for new evaluation features
- Include save, cancel and save+next functionality in edit modal
- Support manual evaluation result modification with reasons
- Add evaluation status badges and feedback display

查看合并请求 AI-PaaS/FastGPT!41
陈志82020 and others added 28 commits September 11, 2025 17:21
[update] Optimize code

查看合并请求 AI-PaaS/FastGPT!72
…rds for evaluation tasks ('eval-task-optimization-code-format' 到 'eval-dev')

feat: add task.dataItem layer API, and optimise the code standards for evaluation tasks

查看合并请求 AI-PaaS/FastGPT!74
… 'eval-dev')

feat: add summary error code and test case

completed
* eval summary error code(评估总结错误码)
* test case(api端到端单测)
* summary config auditlog(总结生成修改配置添加操作日志)

查看合并请求 AI-PaaS/FastGPT!69
- Add getEvalDatasetDataDetail API types and response structure
- Implement authEvaluationDatasetDataReadById authorization function
- Create new API endpoint for fetching evaluation dataset data details
- Enhance dataset list API with optional dataCount for evaluation scene
- Remove obsolete evaluation dataset listDataset.ts file
…sting ('eval-refactor-front' 到 'eval-dev')

feat: Add evaluation dataset data detail and enhance dataset listing

- Add getEvalDatasetDataDetail API types and response structure
- Implement authEvaluationDatasetDataReadById authorization function
- Create new API endpoint for fetching evaluation dataset data details
- Enhance dataset list API with optional dataCount for evaluation scene
- Remove obsolete evaluation dataset listDataset.ts file

查看合并请求 AI-PaaS/FastGPT!75
- Replace dataset and collection selection with new FilesCascader component
- Add multi-language support
- Optimize selection flow with unified cascading selector
- Add "skip knowledge base" option with helpful hints
- Improve UI layout and interaction with consistent modal design
- Add state management for input data modal visibility
…ascader' 到 'eval-dev')

feat: refactor expected answer annotation modal

- Replace dataset and collection selection with new FilesCascader component
- Add multi-language support
- Optimize selection flow with unified cascading selector
- Add "skip knowledge base" option with helpful hints
- Improve UI layout and interaction with consistent modal design
- Add state management for input data modal visibility

查看合并请求 AI-PaaS/FastGPT!76
…asets

- Implement intelligent generation of evaluation datasets
- Add APIs for creating, updating, and deleting evaluation datasets
- Complete the display and manipulation of evaluation dataset data lists
- Implement manual addition and intelligent generation of dataset data
- Add data quality assessment functionality for datasets
- Optimize interaction logic on the dataset details page
…on datasets ('76887/0909/01' 到 'eval-dev')

feat(evaluation): Implement complete functionality for evaluation datasets

- Implement intelligent generation of evaluation datasets
- Add APIs for creating, updating, and deleting evaluation datasets
- Complete the display and manipulation of evaluation dataset data lists
- Implement manual addition and intelligent generation of dataset data
- Add data quality assessment functionality for datasets
- Optimize interaction logic on the dataset details page

查看合并请求 AI-PaaS/FastGPT!67
…k-optimization-validate' 到 'eval-dev')

refactor: optimization param validation of eval-task

Refactor: Optimization of Evaluation Task Parameter Validation

Summary
Unified validation framework for evaluation module with enhanced error handling and async support.

Key Changes
- New validation framework - Added Validatable base class and ValidationResult structure
- Async validation - createEvaluatorInstance and createTargetInstance now support Promise-based validation
- Enhanced error reporting - Detailed error codes, field names, and debug information
- Performance optimization - Optional validation flag for high-performance scenarios
- Improved type safety - Better TypeScript support throughout validation chain

Breaking Changes
- Function signatures changed from sync to async for evaluator/target creation

Files Modified
- packages/global/core/evaluation/validate.ts - NEW validation framework
- packages/service/core/evaluation/evaluator/index.ts - Async evaluator validation
- packages/service/core/evaluation/target/index.ts - Enhanced target validation
- packages/service/core/evaluation/utils/index.ts - Improved parameter validation
- packages/service/core/evaluation/task/processor.ts - Updated instance creation calls

@31202 @64078 @94619

查看合并请求 AI-PaaS/FastGPT!77
- Add evaluation model selector to account model configuration table
- Add multi-language translation support
- Evaluation models will be used for app evaluation and data quality assessment scenarios
- Integrate AIModelSelector component and filter available model list
- Replace mock data with real API calls in task creation modal
- Optimize dimension management with scroll pagination and model selection
- Add multilingual prompts and error handling
- Update task list page with real API integration and status optimization
- Add default model selection utility function
- Improve type definitions and remove unused fields
- Reorganize and expand evaluation error codes with better naming conventions
- Add comprehensive validation constants for name, description, model fields
- Standardize error enum usage across dataset collection and data APIs
- Add detailed field validation with proper length limits and type checking
- Update i18n translations for Chinese, English, and Traditional Chinese
- Enhance API error responses with consistent error enum references
- Update test cases to align with new error handling patterns
…system ('eval-refact-errorcode' 到 'eval-dev')

refactor: Standardize evaluation error handling and validation system

- Reorganize and expand evaluation error codes with better naming conventions
- Add comprehensive validation constants for name, description, model fields
- Standardize error enum usage across dataset collection and data APIs
- Add detailed field validation with proper length limits and type checking
- Update i18n translations for Chinese, English, and Traditional Chinese
- Enhance API error responses with consistent error enum references
- Update test cases to align with new error handling patterns

查看合并请求 AI-PaaS/FastGPT!83
…' 到 'eval-dev')

feat: enhance task creation and management

- Replace mock data with real API calls in task creation modal
- Optimize dimension management with scroll pagination and model selection
- Add multilingual prompts and error handling
- Update task list page with real API integration and status optimization
- Add default model selection utility function
- Improve type definitions and remove unused fields

查看合并请求 AI-PaaS/FastGPT!79
…dels' 到 'eval-dev')

feat: add evaluation model configuration options

- Add evaluation model selector to account model configuration table
- Add multi-language translation support
- Evaluation models will be used for app evaluation and data quality assessment scenarios
- Integrate AIModelSelector component and filter available model list

查看合并请求 AI-PaaS/FastGPT!80
… format ('14864/test-run-score' 到 'eval-dev')

feat: adjust test run result score display to full score of 100 format

查看合并请求 AI-PaaS/FastGPT!81
- Added functionality to navigate from the dataset details page to the file import page, passing the collectionId.
- Refactored the file import page to support two modes:
  - Create a new dataset and import files
  - Append files to an existing dataset
- Optimized the file import form layout and interaction:
  - Dynamically show/hide the dataset name input box based on the mode
  - Improved the file selector UI and error handling
  - Added more detailed template file content
- Implemented file import API call logic:
  - Supports serial import of multiple files
  - Provides detailed error feedback
  - Automatically redirects back to the previous page
- Optimized the automatic evaluation feature:
  - Display the model selector only when enabled
  - Updated prompt text and layout
…87/0909/01' 到 'eval-dev')

feat(evaluation): Enhance the dataset file import feature

- Added functionality to navigate from the dataset details page to the file import page, passing the collectionId.
- Refactored the file import page to support two modes:
  - Create a new dataset and import files
  - Append files to an existing dataset
- Optimized the file import form layout and interaction:
  - Dynamically show/hide the dataset name input box based on the mode
  - Improved the file selector UI and error handling
  - Added more detailed template file content
- Implemented file import API call logic:
  - Supports serial import of multiple files
  - Provides detailed error feedback
  - Automatically redirects back to the previous page
- Optimized the automatic evaluation feature:
  - Display the model selector only when enabled
  - Updated prompt text and layout

查看合并请求 AI-PaaS/FastGPT!78
- Add getEvalDatasetCollectionDetailQuery/Response types for collection details
- Add retryAllTaskBody/Response types for batch retry functionality
- Create shared utils module to eliminate code duplication:
  - Extract getCollectionStatus function
  - Add buildCollectionAggregationPipeline for standardized queries
  - Add formatCollectionBase for consistent response formatting
- Enhance failedTasks endpoint to include datasetId mapping
- Refactor list.ts to use shared utilities
…'eval-dataset-dev' 到 'eval-dev')

feat: add collection detail API and refactor shared utilities

- Add getEvalDatasetCollectionDetailQuery/Response types for collection details
- Add retryAllTaskBody/Response types for batch retry functionality
- Create shared utils module to eliminate code duplication:
  - Extract getCollectionStatus function
  - Add buildCollectionAggregationPipeline for standardized queries
  - Add formatCollectionBase for consistent response formatting
- Enhance failedTasks endpoint to include datasetId mapping
- Refactor list.ts to use shared utilities

查看合并请求 AI-PaaS/FastGPT!86
@github-actions
Copy link

Coverage Report

Status Category Percentage Covered / Total
🔵 Lines 29.24% 20753 / 70962
🔵 Statements 29.24% 20753 / 70962
🔵 Functions 40.64% 604 / 1486
🔵 Branches 74.65% 2227 / 2983
File Coverage
File Stmts Branches Functions Lines Uncovered Lines
Changed Files
packages/global/common/error/errorCode.ts 100% 100% 100% 100%
packages/global/common/error/code/database.ts 100% 100% 100% 100%
packages/global/common/error/code/evaluation.ts 100% 100% 100% 100%
packages/global/common/error/code/system.ts 100% 100% 100% 100%
packages/global/common/error/code/team.ts 100% 100% 100% 100%
packages/global/common/file/constants.ts 100% 100% 100% 100%
packages/global/core/ai/prompt/eval.ts 100% 100% 100% 100%
packages/global/core/chat/constants.ts 100% 100% 100% 100%
packages/global/core/dataset/constants.ts 100% 100% 100% 100%
packages/global/core/dataset/data/constants.ts 95.45% 100% 0% 95.45% 51-52
packages/global/core/evaluation/constants.ts 100% 100% 100% 100%
packages/global/core/evaluation/validate.ts 46.87% 60% 33.33% 46.87% 26-28, 34-36, 52-53, 59-65, 72-73
packages/global/core/evaluation/dataset/constants.ts 100% 100% 100% 100%
packages/global/core/evaluation/metric/constants.ts 100% 100% 100% 100%
packages/global/core/workflow/constants.ts 100% 100% 100% 100%
packages/global/support/permission/constant.ts 100% 100% 100% 100%
packages/global/support/permission/evaluation/constant.ts 100% 100% 100% 100%
packages/global/support/permission/evaluation/controller.ts 82.35% 50% 100% 82.35% 12-14
packages/global/support/permission/user/constant.ts 100% 100% 100% 100%
packages/global/support/permission/user/controller.ts 85.71% 60% 100% 85.71% 25-27, 29-30
packages/global/support/user/audit/constants.ts 100% 100% 100% 100%
packages/service/common/bullmq/index.ts 67.12% 83.33% 50% 67.12% 52-53, 68-91
packages/service/common/file/gridfs/controller.ts 8.9% 100% 0% 8.9% 19-24, 26-192, 195-264
packages/service/common/file/gridfs/schema.ts 100% 100% 100% 100%
packages/service/common/vectorDB/constants.ts 100% 100% 100% 100%
packages/service/common/vectorDB/controller.ts 21.65% 25% 9.09% 21.65% 40, 45-57, 63-88, 91-94, 100-136, 139-176, 179-214
packages/service/common/vectorDB/milvus/index.ts 2.32% 100% 0% 2.32% 21-331
packages/service/common/vectorDB/oceanbase/index.ts 3.14% 100% 0% 3.14% 18-219
packages/service/common/vectorDB/pg/index.ts 6.22% 100% 18.18% 6.22% 24-121, 123-160, 162-195, 197-259, 261-320, 322-338, 340-345, 347-352, 354-369
packages/service/core/ai/model.ts 53.57% 77.77% 27.77% 53.57% 19-20, 23-25, 35-37, 41-43, 47-49, 53-55, 58-65, 67-68
packages/service/core/ai/config/utils.ts 0% 100% 100% 0% 2-269
packages/service/core/app/controller.ts 0% 100% 100% 0% 2-212
packages/service/core/dataset/controller.ts 19.35% 50% 25% 19.35% 21-58, 65-66, 71-127, 130-168
packages/service/core/dataset/schema.ts 98.19% 66.66% 100% 98.19% 188-190
packages/service/core/dataset/collection/controller.ts 8.3% 100% 0% 8.3% 42-312, 316-434
packages/service/core/dataset/collection/schema.ts 98.01% 66.66% 100% 98.01% 189-191
packages/service/core/dataset/collection/utils.ts 7.97% 100% 0% 7.97% 24-61, 64-67, 70-107, 110-133, 136-209, 216-252
packages/service/core/dataset/database/clientManager.ts 0% 100% 100% 0% 2-95
packages/service/core/dataset/database/model/AsyncDB.ts 0% 0% 0% 0% 1-460
packages/service/core/dataset/database/model/dataModel.ts 0% 100% 100% 0% 3-246
packages/service/core/dataset/database/model/mysql.ts 0% 100% 100% 0% 2-33
packages/service/core/dataset/database/model/utils.ts 0% 100% 100% 0% 3-29
packages/service/core/dataset/search/controller.ts 2.94% 100% 0% 2.94% 109-151, 153-942, 950-987, 1009-1093, 1097-1130, 1134-1167, 1171-1238, 1290-1443
packages/service/core/evaluation/common.ts 20.68% 100% 12.5% 20.68% 27-40, 43-46, 49-101, 106-132, 135-149, 152-166, 169-198, 203-224, 227-248, 251-260, 265-276, 278-292, 295-309, 312-335, 388-398, 401-411, 414-422, 425-433, 436-444, 447-461, 464-478
packages/service/core/evaluation/index.ts 0% 0% 0% 0% 1-18
packages/service/core/evaluation/dataset/dataQualityMq.ts 32.81% 0% 0% 32.81% 29, 33-41, 44-47, 50-63, 66-87
packages/service/core/evaluation/dataset/dataQualityProcessor.ts 0% 100% 100% 0% 2-147
packages/service/core/evaluation/dataset/dataSynthesizeMq.ts 0% 0% 0% 0% 1-89
packages/service/core/evaluation/dataset/dataSynthesizeProcessor.ts 0% 100% 100% 0% 2-134
packages/service/core/evaluation/dataset/evalDatasetCollectionSchema.ts 96.61% 50% 100% 96.61% 64-65
packages/service/core/evaluation/dataset/evalDatasetDataSchema.ts 97.95% 66.66% 100% 97.95% 109-110
packages/service/core/evaluation/dataset/smartGenerateMq.ts 0% 0% 0% 0% 1-91
packages/service/core/evaluation/dataset/smartGenerateProcessor.ts 0% 100% 100% 0% 3-198
packages/service/core/evaluation/dataset/utils.ts 83.33% 80% 100% 83.33% 21-37
packages/service/core/evaluation/evaluator/ditingClient.ts 23.63% 100% 50% 23.63% 17-18, 22-70
packages/service/core/evaluation/evaluator/index.ts 51.58% 61.76% 87.5% 51.58% 66-72, 75-80, 83-88, 93-97, 102-111, 117-138, 146-156, 174-214, 244-254, 265-272
packages/service/core/evaluation/metric/schema.ts 96.55% 100% 0% 96.55% 96-98
packages/service/core/evaluation/summary/index.ts 89.62% 70.86% 95% 89.62% 486-490, 519-530, 550-564, 611-613, 733-739, 755-756, 790-798, 809-819, 871-872, 918, 939-945, 1000-1005, 1015
packages/service/core/evaluation/summary/util/weightCalculator.ts 73.58% 53.33% 100% 73.58% 13-14, 21-33, 41, 45-49
packages/service/core/evaluation/synthesizer/ditingSynthesisClient.ts 0% 100% 100% 0% 7-51
packages/service/core/evaluation/synthesizer/index.ts 0% 100% 100% 0% 8-103
packages/service/core/evaluation/target/index.ts 76.51% 65.3% 100% 76.51% 54-63, 118-150, 225-231, 270-280, 295-301, 343-344
packages/service/core/evaluation/task/index.ts 96.1% 72.51% 100% 96.1% 161-162, 199-200, 232-239, 410-411, 431-432, 584-585, 609-610, 618-619, 640-641, 657, 659-660, 695-696, 725-726, 786-788, 842-843, 1253
packages/service/core/evaluation/task/mq.ts 25% 100% 0% 25% 30-32, 35-37, 40-61, 64-85, 88-109, 112-121
packages/service/core/evaluation/task/processor.ts 78.27% 65.97% 88.23% 78.27% 66-67, 77-78, 149-158, 187-188, 234-236, 282-298, 318-320, 375-377, 394-395, 399-400, 403-404, 409-434, 473-486, 499-504, 508-510, 515-520, 544-546, 637-638, 642-650, 684-688, 693-694, 697-698
packages/service/core/evaluation/task/schema.ts 99.48% 50% 100% 99.48% 39
packages/service/core/evaluation/utils/index.ts 43.93% 58.33% 66.66% 43.93% 51-61, 80-90, 93-104, 108-119, 122-132, 136-146, 151-152, 157-167, 177-187, 200-204
packages/service/core/evaluation/utils/jobCleanup.ts 79.9% 79.59% 100% 79.9% 101-107, 136-141, 161-165, 167-172, 228, 236-243, 247-254
packages/service/core/evaluation/utils/tokenLimiter.ts 40.47% 20% 16.66% 40.47% 27-29, 43-49, 73-74, 80-83, 89-90, 96-99, 106-139
packages/service/core/evaluation/utils/usage.ts 100% 100% 100% 100%
packages/service/core/workflow/dispatch/dataset/search.ts 2.97% 100% 0% 2.97% 55-497
packages/service/support/permission/teamLimit.ts 25.4% 20% 30% 25.4% 19-29, 32-45, 59-60, 64-72, 76-95, 108-109, 113-119, 123-130, 133-154, 157-175, 178-199, 202-220
packages/service/support/permission/evaluation/auth.ts 6.94% 100% 0% 6.94% 25-88, 91-124, 128-194, 197-230, 233-270, 274-338, 341-374
packages/service/support/user/audit/util.ts 50% 77.77% 60% 50% 16-22, 25-33, 38-45, 48-52
packages/service/support/wallet/sub/schema.ts 96.05% 0% 0% 96.05% 99-101
packages/service/support/wallet/sub/utils.ts 20.09% 28.57% 20% 20.09% 26-27, 29-30, 35, 56, 62-75, 81-132, 135-213, 216-220, 223-225, 227-242, 245-269
packages/service/support/wallet/usage/controller.ts 7.73% 100% 0% 7.73% 14-27, 30-70, 81-169, 172-197, 200-237, 247-290, 293-335, 338-380, 383-419
packages/web/components/common/Icon/constants.ts 0% 100% 100% 0% 3-515
packages/web/support/user/audit/constants.ts 0% 0% 0% 0% 1-764
projects/app/src/instrumentation.ts 0% 0% 0% 0% 1-82
projects/app/src/pageComponents/dashboard/evaluation/dataset/detail/const.ts 0% 0% 0% 0% 1-43
projects/app/src/pageComponents/dashboard/evaluation/dimension/constants/evaluationTemplates.ts 0% 100% 100% 0% 2-392
projects/app/src/pageComponents/dataset/detail/Import/components/const.ts 0% 0% 0% 0% 1-81
projects/app/src/pageComponents/dataset/detail/Import/components/hooks/useConnectTest.ts 0% 0% 0% 0% 1-64
projects/app/src/pageComponents/dataset/detail/Import/components/hooks/useDataBaseConfig.ts 0% 0% 0% 0% 1-268
projects/app/src/pages/api/common/file/upload.ts 0% 100% 100% 0% 2-154
projects/app/src/pages/api/core/ai/model/updateDefault.ts 0% 100% 100% 0% 2-111
projects/app/src/pages/api/core/dataset/create.ts 95.78% 62.5% 100% 95.78% 64-65, 67-68
projects/app/src/pages/api/core/dataset/detail.ts 0% 0% 0% 0% 1-82
projects/app/src/pages/api/core/dataset/list.ts 0% 0% 0% 0% 1-215
projects/app/src/pages/api/core/dataset/update.ts 0% 0% 0% 0% 1-374
projects/app/src/pages/api/core/dataset/collection/trainingDetail.ts 0% 0% 0% 0% 1-174
projects/app/src/pages/api/core/dataset/database/applyChanges.ts 0% 0% 0% 0% 1-421
projects/app/src/pages/api/core/dataset/database/checkConnection.ts 0% 100% 100% 0% 2-53
projects/app/src/pages/api/core/dataset/database/createCollections.ts 0% 100% 100% 0% 2-154
projects/app/src/pages/api/core/dataset/database/detectChanges.ts 0% 100% 100% 0% 2-244
projects/app/src/pages/api/core/dataset/database/getConfiguration.ts 0% 100% 100% 0% 2-111
projects/app/src/pages/api/core/dataset/database/searchTest.ts 0% 100% 100% 0% 2-135
projects/app/src/pages/api/core/evaluation/dataset/collection/create.ts 100% 97.05% 100% 100%
projects/app/src/pages/api/core/evaluation/dataset/collection/delete.ts 100% 96.15% 100% 100%
projects/app/src/pages/api/core/evaluation/dataset/collection/deleteTask.ts 0% 100% 100% 0% 2-86
projects/app/src/pages/api/core/evaluation/dataset/collection/detail.ts 0% 100% 100% 0% 2-57
projects/app/src/pages/api/core/evaluation/dataset/collection/failedTasks.ts 0% 100% 100% 0% 2-84
projects/app/src/pages/api/core/evaluation/dataset/collection/list.ts 82.4% 76.19% 100% 82.4% 47-49, 65-80, 123-125, 141
projects/app/src/pages/api/core/evaluation/dataset/collection/qualityAssessmentBatch.ts 0% 100% 100% 0% 2-211
projects/app/src/pages/api/core/evaluation/dataset/collection/retryAllTask.ts 0% 100% 100% 0% 2-113
projects/app/src/pages/api/core/evaluation/dataset/collection/retryTask.ts 0% 100% 100% 0% 2-90
projects/app/src/pages/api/core/evaluation/dataset/collection/update.ts 100% 97.56% 100% 100%
projects/app/src/pages/api/core/evaluation/dataset/data/create.ts 98.64% 96.55% 100% 98.64% 95-96
projects/app/src/pages/api/core/evaluation/dataset/data/delete.ts 100% 94.73% 100% 100%
projects/app/src/pages/api/core/evaluation/dataset/data/detail.ts 0% 100% 100% 0% 2-50
projects/app/src/pages/api/core/evaluation/dataset/data/fileId.ts 91.53% 84.15% 100% 91.53% 163-164, 167-168, 171-172, 175-176, 179-180, 215-216, 219-220, 251-252, 265-266, 269-270, 339-340
projects/app/src/pages/api/core/evaluation/dataset/data/list.ts 100% 95.83% 100% 100%
projects/app/src/pages/api/core/evaluation/dataset/data/qualityAssessment.ts 100% 96.29% 100% 100%
projects/app/src/pages/api/core/evaluation/dataset/data/smartGenerate.ts 0% 100% 100% 0% 2-126
projects/app/src/pages/api/core/evaluation/dataset/data/update.ts 96.39% 93.18% 100% 96.39% 95-96, 104-105
projects/app/src/pages/api/core/evaluation/metric/create.ts 88.23% 82.6% 100% 88.23% 34-35, 38-39, 42-43, 50-51
projects/app/src/pages/api/core/evaluation/metric/debug.ts 95.83% 94.64% 100% 95.83% 55-56, 67-68, 71-72
projects/app/src/pages/api/core/evaluation/metric/delete.ts 100% 100% 100% 100%
projects/app/src/pages/api/core/evaluation/metric/detail.ts 90.47% 85.71% 100% 90.47% 17-18
projects/app/src/pages/api/core/evaluation/metric/list.ts 95.04% 77.27% 100% 95.04% 45-47, 104-105, 108
projects/app/src/pages/api/core/evaluation/metric/update.ts 79.1% 80% 100% 79.1% 24-25, 28-29, 32-33, 36-37, 40-41, 44-45, 48-49
projects/app/src/pages/api/core/evaluation/summary/create.ts 100% 100% 100% 100%
projects/app/src/pages/api/core/evaluation/summary/detail.ts 100% 100% 100% 100%
projects/app/src/pages/api/core/evaluation/summary/config/detail.ts 100% 100% 100% 100%
projects/app/src/pages/api/core/evaluation/summary/config/update.ts 80.95% 67.85% 100% 80.95% 42-43, 46-47, 52-53, 57-58, 60-61, 66-67, 69-70, 78-79
projects/app/src/pages/api/core/evaluation/task/create.ts 100% 100% 100% 100%
projects/app/src/pages/api/core/evaluation/task/delete.ts 93.75% 66.66% 100% 93.75% 19-20
projects/app/src/pages/api/core/evaluation/task/detail.ts 100% 100% 100% 100%
projects/app/src/pages/api/core/evaluation/task/list.ts 92.59% 77.77% 100% 92.59% 39-41, 84-86
projects/app/src/pages/api/core/evaluation/task/retryFailed.ts 0% 100% 100% 0% 2-52
projects/app/src/pages/api/core/evaluation/task/start.ts 100% 100% 100% 100%
projects/app/src/pages/api/core/evaluation/task/stats.ts 100% 100% 100% 100%
projects/app/src/pages/api/core/evaluation/task/stop.ts 93.75% 66.66% 100% 93.75% 19-20
projects/app/src/pages/api/core/evaluation/task/update.ts 100% 88.88% 100% 100%
projects/app/src/pages/api/core/evaluation/task/dataItem/delete.ts 100% 100% 100% 100%
projects/app/src/pages/api/core/evaluation/task/dataItem/export.ts 100% 100% 100% 100%
projects/app/src/pages/api/core/evaluation/task/dataItem/list.ts 100% 100% 100% 100%
projects/app/src/pages/api/core/evaluation/task/dataItem/retry.ts 100% 100% 100% 100%
projects/app/src/pages/api/core/evaluation/task/dataItem/update.ts 100% 100% 100% 100%
projects/app/src/pages/api/core/evaluation/task/item/delete.ts 100% 100% 100% 100%
projects/app/src/pages/api/core/evaluation/task/item/detail.ts 100% 100% 100% 100%
projects/app/src/pages/api/core/evaluation/task/item/export.ts 100% 100% 100% 100%
projects/app/src/pages/api/core/evaluation/task/item/list.ts 100% 100% 100% 100%
projects/app/src/pages/api/core/evaluation/task/item/retry.ts 100% 100% 100% 100%
projects/app/src/pages/api/core/evaluation/task/item/update.ts 95.45% 83.33% 100% 95.45% 32-33
projects/app/src/pages/api/support/user/team/limit/evalDatasetDataLimit.ts 0% 100% 100% 0% 2-19
projects/app/src/pages/api/support/user/team/limit/evalDatasetLimit.ts 0% 100% 100% 0% 2-19
projects/app/src/pages/api/support/user/team/limit/evalMetricLimit.ts 0% 100% 100% 0% 2-19
projects/app/src/pages/api/support/user/team/limit/evaluationTaskLimit.ts 0% 100% 100% 0% 2-19
projects/app/src/pages/api/support/user/team/plan/getTeamPlanStatus.ts 0% 100% 100% 0% 2-77
projects/app/src/service/common/system/index.ts 22.5% 100% 15.38% 22.5% 26-46, 52-54, 57-58, 61-63, 66-68, 71-73, 83-103, 126-249
projects/app/src/service/core/dataset/queues/generateDatabaseSchemaVector.ts 0% 0% 0% 0% 1-459
projects/app/src/service/core/dataset/training/utils.ts 0% 0% 0% 0% 1-40
projects/app/src/service/support/wallet/usage/push.ts 0% 0% 0% 0% 1-322
projects/app/src/web/common/file/controller.ts 0% 0% 0% 0% 1-88
projects/app/src/web/common/system/utils.ts 0% 100% 100% 0% 5-83
projects/app/src/web/core/evaluation/dataset.ts 0% 0% 0% 0% 1-106
projects/app/src/web/core/evaluation/dimension.ts 0% 0% 0% 0% 1-75
projects/app/src/web/core/evaluation/task.ts 0% 0% 0% 0% 1-230
projects/app/src/web/support/user/team/api.ts 0% 0% 0% 0% 1-125
Generated in workflow #18 for commit 28d1771 by the Vitest Coverage Report Action

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

7 participants