from typing import List, Optional
from tea_client.models import TeaClientModel
from paperswithcode.models.page import Page
[docs]class EvaluationTable(TeaClientModel):
"""Evaluation table object.
Attributes:
id (str): Evaluation table ID.
task (str): ID of the task used in evaluation.
dataset (str): ID of the dataset used in evaluation.
description (str): Evaluation table description.
mirror_url (str, optional): URL to the evaluation table that this table
is a mirror of.
"""
id: str
task: str
dataset: str
description: str = ""
mirror_url: Optional[str] = None
[docs]class EvaluationTableCreateRequest(TeaClientModel):
"""Evaluation table create request object.
Attributes:
task (str): ID of the task used in evaluation.
dataset (str): ID of the dataset used in evaluation.
description (str): Evaluation table description.
mirror_url (str, optional): URL to the evaluation table that this table
is a mirror of.
"""
task: str
dataset: str
description: str = ""
mirror_url: Optional[str] = None
[docs]class EvaluationTableUpdateRequest(TeaClientModel):
"""Evaluation table update request object.
Attributes:
task (str, optional): ID of the task used in evaluation.
dataset (str, optional): ID of the dataset used in evaluation.
description (str, optional): Evaluation table description.
mirror_url (str, optional): URL to the evaluation table that this table
is a mirror of.f
"""
task: Optional[str] = None
dataset: Optional[str] = None
description: Optional[str] = None
mirror_url: Optional[str] = None
[docs]class EvaluationTables(Page):
"""Object representing a paginated page of evaluation tables.
Attributes:
count (int): Number of elements matching the query.
next_page (int, optional): Number of the next page.
previous_page (int, optional): Number of the previous page.
results (List[SotaPartial]): List of evaluation tables on this page.
"""
results: List[EvaluationTable]