Carcass wrote:
A driver completed the first 20 miles of a 40-mile trip at an average speed of 50 miles per hour. At what average speed must the driver complete the remaining 20 miles to achieve an average speed of 60 miles per hour for the entire 40-mile trip? (Assume that the driver did not make any stops during the 40-mile trip.)
(A) 65 mph
(B) 68 mph
(C) 70 mph
(D) 75 mph
(E) 80 mph
The total distance is 40 miles, and we want the average speed to be 60 miles per hour.
Average speed = (total distance)/(total time)
So, we get: 60 = (40 miles)/(total time)
Solve equation to get: total time =
2/3 hoursSo, the TIME for the ENTIRE 40-mile trip needs to be
2/3 hours. driver completed the first 20 miles of a 40-mile trip at an average speed of 50 miles per hour.How much time was spent on this FIRST PART of the trip?
time = distance/speed
So, time = 20/50 =
2/5 hoursThe ENTIRE trip needs to be
2/3 hours, and the FIRST PART of the trip took
2/5 hours2/3 hours -
2/5 hours =
10/15 hours -
6/15 hours=
4/15 hoursSo, the SECOND PART of the trip needs to take
4/15 hoursThe SECOND PART of the trip is 20 miles, and the time is
4/15 hoursSpeed = distance/time
So, speed = 20/(
4/15)
= (20)(15/4)
= 75
Answer: D