graph LR
src_lighteval_pipeline_Pipeline["src.lighteval.pipeline.Pipeline"]
src_lighteval_logging_evaluation_tracker_EvaluationTracker["src.lighteval.logging.evaluation_tracker.EvaluationTracker"]
src_lighteval_pipeline_Pipeline -- "delegates results to" --> src_lighteval_logging_evaluation_tracker_EvaluationTracker
This subsystem is responsible for the persistence, loading, and presentation of evaluation results, ensuring that detailed responses and aggregated scores are managed effectively and can be integrated with external platforms for visualization and sharing.
This component orchestrates the final stages of the evaluation pipeline, focusing on preparing, computing, and initiating the display and saving of results. It acts as the high-level controller for ensuring results are ready for output and presentation. Its responsibilities include post-processing outputs, computing final metrics, and triggering the persistence and reporting mechanisms.
Related Classes/Methods:
src.lighteval.pipeline.Pipeline:_post_process_outputssrc.lighteval.pipeline.Pipeline:_compute_metricssrc.lighteval.pipeline.Pipeline:save_and_push_resultssrc.lighteval.pipeline.Pipeline:show_resultssrc.lighteval.pipeline.Pipeline:get_results
This component manages the actual storage, loading, and external integration of evaluation results and detailed responses. It handles the low-level persistence mechanisms (e.g., saving to disk) and interfaces with various external reporting platforms such as Hugging Face Hub, TensorBoard, and Weights & Biases. It ensures the long-term availability and shareability of evaluation outcomes.
Related Classes/Methods:
src.lighteval.logging.evaluation_tracker.EvaluationTracker:savesrc.lighteval.logging.evaluation_tracker.EvaluationTracker:save_detailssrc.lighteval.logging.evaluation_tracker.EvaluationTracker:save_resultssrc.lighteval.logging.evaluation_tracker.EvaluationTracker:push_to_tensorboardsrc.lighteval.logging.evaluation_tracker.EvaluationTracker:push_to_wandbsrc.lighteval.logging.evaluation_tracker.EvaluationTracker:push_to_hubsrc.lighteval.logging.evaluation_tracker.EvaluationTracker:load_details_datasetssrc.lighteval.logging.evaluation_tracker.EvaluationTracker:recreate_metadata_card