- #1
DKATyler
- 4
- 0
I am currently attempting to solve the following problem for a calculator I am constructing in Java.
Company 1 pays A% Interest every T days
Company 2 pays B% Interest (>A%) every T days
Company 1 has an initial investment of $X dollars
Company 2 requires an inital fee of $Y dollars to invest in their company ($Y does not pay any interest! It's a fee.)
At what point T in terms of A, B, X and Y should I stop investing in Company 1 and begin investing in Company 2 to obtain the maximum return at T=infinity? Investments are non-refundable.
Example:
$1,000 invested in A at 3% interest
B pays 8% interest
B requires a $1,000 fee to invest.
By brute force calculation, the optimum T value is approx 16 for this example. ($1604.71 invested in Company 1). This results in the first interest earning investment in B at T=39. (Only slightly later then T=0 with $1000 in company where the first interest earning payment would be made at T=36).
Company 1 pays A% Interest every T days
Company 2 pays B% Interest (>A%) every T days
Company 1 has an initial investment of $X dollars
Company 2 requires an inital fee of $Y dollars to invest in their company ($Y does not pay any interest! It's a fee.)
At what point T in terms of A, B, X and Y should I stop investing in Company 1 and begin investing in Company 2 to obtain the maximum return at T=infinity? Investments are non-refundable.
Example:
$1,000 invested in A at 3% interest
B pays 8% interest
B requires a $1,000 fee to invest.
By brute force calculation, the optimum T value is approx 16 for this example. ($1604.71 invested in Company 1). This results in the first interest earning investment in B at T=39. (Only slightly later then T=0 with $1000 in company where the first interest earning payment would be made at T=36).