Here is my problem: I have to use a voltage source that supplies a load through a resistor. The voltage source has a rating of 1000 volts and a power rating of 10000 watts. The resistor is 5 ohms with a power ratings of 118.975 watts. The resistance of the load may vary randomly between 200 and 600 ohms. Using 100000 samples calculate the number of times the power rating of the resistor is exceeded, average power delivered by source, maximum power delivered by source, and the minimum power delivered by source. The output of the program should consist of 4 numbers, each number being output on a separate line in the order listed above. The numbers should be output with a precision of 2. I hoping that someone can help me out and explain to me what i need to do.