- #1

- 4

- 0

## Main Question or Discussion Point

I am currently attempting to solve the following problem for a calculator I am constructing in Java.

Company 1 pays A% Interest every T days

Company 2 pays B% Interest (>A%) every T days

Company 1 has an initial investment of $X dollars

Company 2 requires an inital fee of $Y dollars to invest in their company ($Y does not pay any interest! It's a fee.)

At what point T in terms of A, B, X and Y should I stop investing in Company 1 and begin investing in Company 2 to obtain the maximum return at T=infinity? Investments are non-refundable.

Example:

$1,000 invested in A at 3% interest

B pays 8% interest

B requires a $1,000 fee to invest.

By brute force calculation, the optimum T value is approx 16 for this example. ($1604.71 invested in Company 1). This results in the first interest earning investment in B at T=39. (Only slightly later then T=0 with $1000 in company where the first interest earning payment would be made at T=36).

Company 1 pays A% Interest every T days

Company 2 pays B% Interest (>A%) every T days

Company 1 has an initial investment of $X dollars

Company 2 requires an inital fee of $Y dollars to invest in their company ($Y does not pay any interest! It's a fee.)

At what point T in terms of A, B, X and Y should I stop investing in Company 1 and begin investing in Company 2 to obtain the maximum return at T=infinity? Investments are non-refundable.

Example:

$1,000 invested in A at 3% interest

B pays 8% interest

B requires a $1,000 fee to invest.

By brute force calculation, the optimum T value is approx 16 for this example. ($1604.71 invested in Company 1). This results in the first interest earning investment in B at T=39. (Only slightly later then T=0 with $1000 in company where the first interest earning payment would be made at T=36).