asked 106k views
4 votes
Roger ran 3.2 miles the first day of the fundraiser, 4.0 miles the second day, and 5.1 the last day. If he earned $0.15 per foot for charity, how much did he earn?

asked
User AndreasW
by
8.9k points

1 Answer

4 votes
3.2 + 4 + 5.1 = 12.3, we then divide the total amount of miles ran by the cost per mile, 0.15. We divide because it's a decimal. The answer is 82 dollars.
answered
User ChenZ
by
8.5k points

No related questions found

Welcome to Qamnty — a place to ask, share, and grow together. Join our community and get real answers from real people.