Evaluation Table Models

class paperswithcode.models.evaluation.Metric(*, id: str, name: str, description: str, is_loss: bool)[source]

Metric object.

Metric used for evaluation.

id

Metric id.

Type

str

name

Metric name.

Type

str

description

Metric description.

Type

str

is_loss

Is this a loss metric.

Type

bool

Create a new model by parsing and validating input data from keyword arguments.

Raises ValidationError if the input data cannot be parsed to form a valid model.

class paperswithcode.models.evaluation.Metrics(*, count: int, next_page: int = None, previous_page: int = None, results: List[paperswithcode.models.evaluation.metric.Metric])[source]

Object representing a paginated page of metrics.

count

Number of elements matching the query.

Type

int

next_page

Number of the next page.

Type

int, optional

previous_page

Number of the previous page.

Type

int, optional

results

List of metrics on this page.

Type

List[Metric]

Create a new model by parsing and validating input data from keyword arguments.

Raises ValidationError if the input data cannot be parsed to form a valid model.

class paperswithcode.models.evaluation.MetricCreateRequest(*, name: str, description: str, is_loss: bool)[source]

Metric object.

Metric used for evaluation.

name

Metric name.

Type

str

description

Metric description.

Type

str

is_loss

Is this a loss metric.

Type

bool

Create a new model by parsing and validating input data from keyword arguments.

Raises ValidationError if the input data cannot be parsed to form a valid model.

class paperswithcode.models.evaluation.MetricUpdateRequest(*, name: str = None, description: str = None, is_loss: bool = None)[source]

Metric object.

Metric used for evaluation.

name

Metric name.

Type

str, optional

description

Metric description.

Type

str, optional

is_loss

Is this a loss metric.

Type

bool, optional

Create a new model by parsing and validating input data from keyword arguments.

Raises ValidationError if the input data cannot be parsed to form a valid model.

class paperswithcode.models.evaluation.Result(*, id: str, best_rank: int = None, metrics: dict, methodology: str, uses_additional_data: bool, paper: str = None, best_metric: str = None, evaluated_on: str = None, external_source_url: str = None)[source]

Evaluation table row object.

id

Result id.

Type

str

best_rank

Best rank of the row.

Type

int, optional

metrics

Dictionary of metrics and metric values.

Type

dict

methodology

Methodology used for this implementation.

Type

str

uses_additional_data

Does this evaluation uses additional data not provided in the dataset used for other evaluations.

Type

bool

paper

Paper describing the evaluation.

Type

str, optional

best_metric

Name of the best metric.

Type

str, optional

evaluated_on

Date of the result evaluation in YYYY-MM-DD format.

Type

str, optional

external_source_url

The URL to the external source (eg competition).

Type

str, option

Create a new model by parsing and validating input data from keyword arguments.

Raises ValidationError if the input data cannot be parsed to form a valid model.

class paperswithcode.models.evaluation.Results(*, count: int, next_page: int = None, previous_page: int = None, results: List[paperswithcode.models.evaluation.result.Result])[source]

Object representing a paginated page of results.

count

Number of elements matching the query.

Type

int

next_page

Number of the next page.

Type

int, optional

previous_page

Number of the previous page.

Type

int, optional

results

List of results on this page.

Type

List[Result]

Create a new model by parsing and validating input data from keyword arguments.

Raises ValidationError if the input data cannot be parsed to form a valid model.

class paperswithcode.models.evaluation.ResultCreateRequest(*, metrics: dict, methodology: str, uses_additional_data: bool = False, paper: str = None, evaluated_on: str = None, external_source_url: str = None)[source]

Evaluation table row object.

metrics

Dictionary of metrics and metric values.

Type

dict

methodology

Methodology used for this implementation.

Type

str

uses_additional_data

Does this evaluation uses additional data not provided in the dataset used for other valuations.

Type

bool, optional

paper

Paper describing the evaluation.

Type

str, optional

evaluated_on

Date of the result evaluation: YYYY-MM-DD format.

Type

str, optional

external_source_url

The URL to the external source (eg competition).

Type

str, option

Create a new model by parsing and validating input data from keyword arguments.

Raises ValidationError if the input data cannot be parsed to form a valid model.

class paperswithcode.models.evaluation.ResultUpdateRequest(*, metrics: dict = None, methodology: str = None, uses_additional_data: bool = None, paper: str = None, evaluated_on: str = None, external_source_url: str = None)[source]

Evaluation table row object.

metrics

Dictionary of metrics and metric values.

Type

dict, optional

methodology

Methodology used for this implementation.

Type

str, optional

uses_additional_data

Does this evaluation uses additional data not provided in the dataset used for other evaluations.

Type

bool, optional

paper

Paper describing the evaluation.

Type

str, optional

evaluated_on

Date of the result evaluation: YYYY-MM-DD format.

Type

datetime, optional

external_source_url

The URL to the external source (eg competition).

Type

str, option

Create a new model by parsing and validating input data from keyword arguments.

Raises ValidationError if the input data cannot be parsed to form a valid model.

class paperswithcode.models.evaluation.EvaluationTable(*, id: str, task: str, dataset: str, description: str = '', mirror_url: str = None)[source]

Evaluation table object.

id

Evaluation table ID.

Type

str

task

ID of the task used in evaluation.

Type

str

dataset

ID of the dataset used in evaluation.

Type

str

description

Evaluation table description.

Type

str

mirror_url

URL to the evaluation table that this table is a mirror of.

Type

str, optional

Create a new model by parsing and validating input data from keyword arguments.

Raises ValidationError if the input data cannot be parsed to form a valid model.

class paperswithcode.models.evaluation.EvaluationTables(*, count: int, next_page: int = None, previous_page: int = None, results: List[paperswithcode.models.evaluation.table.EvaluationTable])[source]

Object representing a paginated page of evaluation tables.

count

Number of elements matching the query.

Type

int

next_page

Number of the next page.

Type

int, optional

previous_page

Number of the previous page.

Type

int, optional

results

List of evaluation tables on this page.

Type

List[SotaPartial]

Create a new model by parsing and validating input data from keyword arguments.

Raises ValidationError if the input data cannot be parsed to form a valid model.

class paperswithcode.models.evaluation.EvaluationTableCreateRequest(*, task: str, dataset: str, description: str = '', mirror_url: str = None)[source]

Evaluation table create request object.

task

ID of the task used in evaluation.

Type

str

dataset

ID of the dataset used in evaluation.

Type

str

description

Evaluation table description.

Type

str

mirror_url

URL to the evaluation table that this table is a mirror of.

Type

str, optional

Create a new model by parsing and validating input data from keyword arguments.

Raises ValidationError if the input data cannot be parsed to form a valid model.

class paperswithcode.models.evaluation.EvaluationTableUpdateRequest(*, task: str = None, dataset: str = None, description: str = None, mirror_url: str = None)[source]

Evaluation table update request object.

task

ID of the task used in evaluation.

Type

str, optional

dataset

ID of the dataset used in evaluation.

Type

str, optional

description

Evaluation table description.

Type

str, optional

mirror_url

URL to the evaluation table that this table is a mirror of.f

Type

str, optional

Create a new model by parsing and validating input data from keyword arguments.

Raises ValidationError if the input data cannot be parsed to form a valid model.

class paperswithcode.models.evaluation.ResultSyncRequest(*, metrics: dict, methodology: str, paper: str = None, uses_additional_data: bool = False, external_id: str = '', evaluated_on: str, external_source_url: str = None)[source]

Evaluation table row object.

metrics

Dictionary of metrics and metric values.

Type

dict

methodology

Methodology used for this implementation.

Type

str

uses_additional_data

Does this evaluation uses additional data not provided in the dataset used for other evaluations.

Type

bool

paper

Paper describing the evaluation.

Type

str, optional

external_id

Optional external ID used to identify rows when doing sync.

Type

str, optional

evaluated_on

Evaluation date in YYYY-MM-DD format

Type

str

external_source_url

The URL to the external source (eg competition).

Type

str, option

Create a new model by parsing and validating input data from keyword arguments.

Raises ValidationError if the input data cannot be parsed to form a valid model.

class paperswithcode.models.evaluation.MetricSyncRequest(*, name: str, description: str = '', is_loss: bool = True)[source]

Metric object.

Metric used for evaluation.

name

Metric name.

Type

str

description

Metric description.

Type

str

is_loss

Is this a loss metric.

Type

bool

Create a new model by parsing and validating input data from keyword arguments.

Raises ValidationError if the input data cannot be parsed to form a valid model.

class paperswithcode.models.evaluation.EvaluationTableSyncRequest(*, task: str, dataset: str, description: str = '', mirror_url: str = None, external_id: str = None, metrics: List[paperswithcode.models.evaluation.synchronize.MetricSyncRequest] = None, results: List[paperswithcode.models.evaluation.synchronize.ResultSyncRequest] = None)[source]

Evaluation table object.

task

ID of the task used in evaluation.

Type

str

dataset

ID of the dataset used in evaluation.

Type

str

description

Evaluation table description.

Type

str

mirror_url

URL to the evaluation table that this table is a mirror of.

Type

str, optional

external_id

Optional external ID used to identify rows when doing sync.

Type

str, optional

metric

List of MetricSyncRequest objects used in the evaluation.

Type

list

results

List of ResultSyncRequest objects - results of the evaluation.

Type

list

Create a new model by parsing and validating input data from keyword arguments.

Raises ValidationError if the input data cannot be parsed to form a valid model.

class paperswithcode.models.evaluation.ResultSyncResponse(*, id: str, metrics: dict, methodology: str, paper: str = None, uses_additional_data: bool = False, external_id: str = '', evaluated_on: str = None, external_source_url: str = None)[source]

Evaluation table row object.

id

Result id.

Type

str

metrics

Dictionary of metrics and metric values.

Type

dict

methodology

Methodology used for this implementation.

Type

str

uses_additional_data

Does this evaluation uses additional data not provided in the dataset used for other evaluations.

Type

bool

paper

Paper describing the evaluation.

Type

str, optional

external_id

Optional external ID used to identify rows when doing sync.

Type

str, optional

evaluated_on

Evaluation date in YYYY-MM-DD format

Type

str, optional

external_source_url

The URL to the external source (eg competition)

Type

str, option

Create a new model by parsing and validating input data from keyword arguments.

Raises ValidationError if the input data cannot be parsed to form a valid model.

class paperswithcode.models.evaluation.MetricSyncResponse(*, name: str, description: str = '', is_loss: bool = True)[source]

Metric object.

Metric used for evaluation.

name

Metric name.

Type

str

description

Metric description.

Type

str

is_loss

Is this a loss metric.

Type

bool

Create a new model by parsing and validating input data from keyword arguments.

Raises ValidationError if the input data cannot be parsed to form a valid model.

class paperswithcode.models.evaluation.EvaluationTableSyncResponse(*, id: str, task: str, dataset: str, description: str = '', mirror_url: str = None, external_id: str = '', metrics: List[paperswithcode.models.evaluation.synchronize.MetricSyncResponse] = None, results: List[paperswithcode.models.evaluation.synchronize.ResultSyncResponse] = None)[source]

Evaluation table object.

id

Evaluation table ID.

Type

str

task

ID of the task used in evaluation.

Type

str

dataset

ID of the dataset used in evaluation.

Type

str

description

Evaluation table description.

Type

str

mirror_url

URL to the evaluation table that this table is a mirror of.

Type

str, optional

external_id

Optional external ID used to identify rows when doing sync.

Type

str, optional

metric

List of metrics sync objects used in the evaluation.

Type

list

results

List of result sync objects - results of the evaluation.

Type

list

Create a new model by parsing and validating input data from keyword arguments.

Raises ValidationError if the input data cannot be parsed to form a valid model.