githubEdit

SLAs and QoS Metrics

The Service Level Agreements (SLAs) and Quality of Service (QoS) metrics are used to evaluate and monitor providers.

circle-check

These metrics ensure a consistent and reliable experience for consumers while providing transparency and accountability for providers.

SLA Definitions

The following SLA parameters define the minimum standards providers must meet. The ISP Market asks consumers to define target SLA values when allocating a Data Unit. Providers can accept or decline allocations based on these terms. Conversely, providers can publish Data Units with predefined SLA profiles that consumers can select directly.

Key Metrics
Definitions

Latency

The time it takes to complete a read or write operation.

Example: Maximum average latency of 50ms for read/write operations

Upload Speed

The speed at which data can be uploaded to the storage service.

Example: Minimum speed of 10 MB/s for uploads of files 1 GB or smaller.

Download Speed

The speed at which data can be downloaded from the storage service.

Example: Minimum speed of 20 MB/s for downloads of files 1 GB or smaller.

Uptime

The percentage of time the storage service is operational and accessible.

Example: 99.95% uptime over a rolling 30-day period.

Error Rate

The proportion of failed operations (e.g., upload, download, or delete) compared to the total operations.

Example: Less than 0.01% failed operations per month.


QoS Records

The QoS records serve as the backbone for evaluating SLAs. These records are continuously generated through monitoring and probing mechanisms.

Structure of QoS Records

Each QoS record consists of the following fields:

  • Timestamp: The exact time when the record was created.

  • ID: A unique identifier for the provider or the consumer.

  • Operation Type: The type of operation being measured (e.g., read, write, upload, download).

  • Latency: Measured latency for the operation.

  • Throughput: Upload or download speed, depending on the operation type.

  • Success Status: A flag indicating whether the operation succeeded or failed.

  • Error Details (if applicable): A description of the error, if the operation failed.

Data Collection Process

QoS data is collected through a combination of:

  1. Probing Operations: Automated test operations reported to the platform where the consumers and providers must report their QoS records. The report will support custom period of reports but it must be equal of lower than 24 hours between reports.

  2. System Metrics: Real-time monitoring of uptime and operational logs.

High-Level Workflow

  1. Probing: Scheduled reports that contains metrics such as read, write, upload, and download operations at regular intervals.

  2. Measurement: Each operation records its latency, throughput, and success status from the consumer and providers.

  3. Aggregation: QoS records are aggregated over time to calculate averages, percentages, and trends.

  4. Validation: Aggregated data is validated against SLA requirements found in the smart contract to determine compliance.

  5. Storage: All QoS records are securely stored in a time-series database for long-term analysis.

Reporting

Summarized QoS rolling data is sent to the Smart Contract for on-chain enforcement and platform visibility.

Apart from that, other external mechanisms can be implemented to generate and send regular SLA compliance reports to Providers based on custom periods.

The summarized data includes:

  • Average latency, upload speed, and download speed over the reporting period.

  • Uptime percentage.

  • Error rate statistics.

Enforcement

Failure to meet SLA requirements may result in penalties or reduced reputation scores. Consumers and providers are encouraged to consistently monitor their performance and address any deficiencies proactively.

Last updated

Was this helpful?