Hide

Problem D
Stream Lag

/problems/streamlag/file/statement/en/img-0001.png
Image from freesvg.org

Live stream audiences often encounter undesired stream lag. The lag may occur for multiple reasons, such as slow network speed, high stream resolution, not enough processing power on the client hardware, etc. In this problem, we will model and compute the stream lag for one hypothetical stream.

The stream has a video content that is segmented into $n$ network packets to send, numbered from $1$ to $n$. Each packet contains a small segment of the streamed video with a length of exactly one second. A stream lag is a time period in which the stream audience is not watching any content while waiting for the stream packets to arrive. Ideally, a stream audience experiencing zero lag shall receive packet $i$ at the beginning of the $i$th second, in which case the audience can seamlessly watch the entire streamed video.

In reality, any of the $n$ packets may be received at any moment, and not necessarily in order from $1$ to $n$. A stream client will only start playing packet $i$ if it has played all its preceding packets. If this is not the case, the stream client will wait until all packets before $i$ have been received and played. The stream client keeps all received packets that cannot yet be played in its buffer and is able to retrieve them in no time when they are ready to be played. If a packet is not available when the time to play it arrives, the stream lags and viewers fall behind from the live stream. The stream client plays each packet for exactly one second at its original speed. All packets are to be played even when the play time lags much behind the live stream.

Given the arrival time of the $n$ packets in chronological order, compute the total lag time that a stream audience will experience.

Input

The first line of input has a single integer $n$ ($1 \leq n \leq 1\, 000$). This is followed by $n$ lines. The $i$th line has two integers $t_ i$ and $i$ ($1 \leq t_ i \leq 10^9, 1 \leq i \leq n$), which means that packet $i$ arrives at the beginning of the $t_ i$th second. The values of $t_ i$’s are non-decreasing. Packets may arrive at the same time.

Output

Output the total lag time based on the given packet arrival time.

Sample Input 1 Sample Output 1
5
1 1
3 2
4 5
4 3
5 4
1
Sample Input 2 Sample Output 2
4
1 1
3 3
4 2
8 4
4
Sample Input 3 Sample Output 3
3
1 1
2 2
2 3
0

Please log in to submit a solution to this problem

Log in