Hello all - I have to create a program in C - I have very less experience in C. Could somebody help me in writing this program?

Here are the details:-

Suppose N stations are waiting for another packet to finish on an Ethernet.

All transmit at once when the other packet is finished, and collide. Write a

program to simulate the continuation of these attempts, and to determine how

long it takes before one succeeds. Make the following simplifications: ignore

interframe spacing, ignore variability in collision times (so that

retransmission is always after an exact integral multiple of the 51.2microsec

slot time), and assume that each collision uses up exactly one slot time.

Model the time, T, in units of slot times, so a collision at time T followed

by a backoff of k=0 would result in a retransmission attempt at time T+1.

Find out the average delay for N=20, N=40, N=100, and maybe some larger values

too. Do your data support the notion that the delay is linear in N? (It should.)

I appreciate any guidlines or help.

Thanks