Source code for paperswithcode.models.evaluation.result

from datetime import datetime
from typing import Optional, List

from tea_client.models import TeaClientModel

from paperswithcode.models.page import Page


[docs]class Result(TeaClientModel): """Evaluation table row object. Attributes: id (str): Result id. best_rank (int, optional): Best rank of the row. metrics (dict): Dictionary of metrics and metric values. methodology (str): Methodology used for this implementation. uses_additional_data (bool): Does this evaluation uses additional data not provided in the dataset used for other evaluations. paper (str, optional): Paper describing the evaluation. best_metric (str, optional): Name of the best metric. evaluated_on (str, optional): Date of the result evaluation in YYYY-MM-DD format. external_source_url (str, option): The URL to the external source (eg competition). """ id: str best_rank: Optional[int] metrics: dict methodology: str uses_additional_data: bool paper: Optional[str] best_metric: Optional[str] evaluated_on: Optional[str] external_source_url: Optional[str]
[docs]class Results(Page): """Object representing a paginated page of results. Attributes: count (int): Number of elements matching the query. next_page (int, optional): Number of the next page. previous_page (int, optional): Number of the previous page. results (List[Result]): List of results on this page. """ results: List[Result]
class _ResultRequest(TeaClientModel): def dict( self, *, include=None, exclude=None, by_alias: bool = False, skip_defaults: bool = None, exclude_unset: bool = False, exclude_defaults: bool = False, exclude_none: bool = False, ): d = super().dict( include=include, exclude=exclude, by_alias=by_alias, skip_defaults=skip_defaults, exclude_unset=exclude_unset, exclude_defaults=exclude_defaults, exclude_none=exclude_none, ) evaluated_on = d.get("evaluated_on") if isinstance(evaluated_on, datetime): d["evaluated_on"] = evaluated_on.strftime("%Y-%m-%d") return d
[docs]class ResultCreateRequest(_ResultRequest): """Evaluation table row object. Attributes: metrics (dict): Dictionary of metrics and metric values. methodology (str): Methodology used for this implementation. uses_additional_data (bool, optional): Does this evaluation uses additional data not provided in the dataset used for other valuations. paper (str, optional): Paper describing the evaluation. evaluated_on (str, optional): Date of the result evaluation: YYYY-MM-DD format. external_source_url (str, option): The URL to the external source (eg competition). """ metrics: dict methodology: str uses_additional_data: Optional[bool] = False paper: Optional[str] = None evaluated_on: Optional[str] = None external_source_url: Optional[str] = None
[docs]class ResultUpdateRequest(_ResultRequest): """Evaluation table row object. Attributes: metrics (dict, optional): Dictionary of metrics and metric values. methodology (str, optional): Methodology used for this implementation. uses_additional_data (bool, optional): Does this evaluation uses additional data not provided in the dataset used for other evaluations. paper (str, optional): Paper describing the evaluation. evaluated_on (datetime, optional): Date of the result evaluation: YYYY-MM-DD format. external_source_url (str, option): The URL to the external source (eg competition). """ metrics: Optional[dict] = None methodology: Optional[str] = None uses_additional_data: Optional[bool] = None paper: Optional[str] = None evaluated_on: Optional[str] = None external_source_url: Optional[str] = None