Abstract:
In one embodiment, a method includes measuring at a cache node located on a data path between a content source and an adaptive streaming client, a time to retrieve content from the content source, and rate limiting streaming of the content from the cache node to the adaptive streaming client based on the measured time to retrieve the content. An apparatus and logic are also disclosed herein.
Abstract:
In one embodiment, a method includes estimating a current queuing latency, the estimated current queuing latency being associated with a queue of packets maintained in a buffer. The method also includes calculating a current drop or mark probability, the current drop or mark probability being associated with a probability that packets associated with the queue of packets will be dropped or marked. A rate at which the packets associated with the queue of packets are dequeued from the buffer is estimated in order to estimate the current queuing latency. The current drop or mark probability is calculated using the current estimated queuing latency.
Abstract:
In one embodiment, a method includes estimating a current queuing latency, the estimated current queuing latency being associated with a queue of packets maintained in a buffer. The method also includes calculating a current drop or mark probability, the current drop or mark probability being associated with a probability that packets associated with the queue of packets will be dropped or marked. A rate at which the packets associated with the queue of packets are dequeued from the buffer is estimated in order to estimate the current queuing latency. The current drop or mark probability is calculated using the current estimated queuing latency.