Nakita rode her skateboard at an average speed of 8.8 feet per second for 0.2 miles. She wants to calculate how long it took her to skateboard that distance. Nakita uses the formula d=rt (d = distance, r = rate, t = time), and she knows that 1 mile = 5280 feet and that 1 minute = 60 seconds. How many minutes did it take Nakita to skateboard 0.2 miles at her average speed of 8.8 ft/sec?

Respuesta :

Answer: [tex]2\ minutes[/tex]

Step-by-step explanation:

We know that:

[tex]d=rt[/tex]

Where "d" is distance, "r" is rate and "t" is time.

Solved for the time "t":

[tex]t=\frac{d}{r}[/tex]

The first step is to convert the distance from miles to feet.

Since [tex]1\ mile=5,280\ feet[/tex], we get:

[tex]d=(0.2\ mi)(\frac{5,280\ ft}{1\ mi})\\\\d=1,056\ ft[/tex]

Knowing that:

[tex]r=8.8\ \frac{ft}{s}[/tex]

We can substitute values into [tex]t=\frac{d}{r}[/tex] in order to find the time in seconds:

[tex]t=\frac{1,056\ ft}{8.8\ \frac{ft}{s}}\\\\t=120\ s[/tex]

Since:

[tex]1\ minute = 60\ seconds[/tex]

The time in minutes is:

[tex]t=(120\ s)(\frac{1\ min}{60\ s})\\\\t=2\ min[/tex]

Q&A Education