vanhees71 said:
The phase space (either of a single particle, often called ##\mu## space as you've explained above or of the entire system of ##N## particles, called ##\Gamma## space, which I think goes back to the famous article by Ehrenfest&Ehrenfest) is a continuous space of states, and thus the probabilities to find a particle or the system of particles in a specific configuration is discribed by a probability distribution, i.e., a function with dimension ##\text{action}^3## or ##\text{action}^{3N}##. In classical physics the only place where the action occurs is in Hamilton's principle of least action, but there is no specific natural measure of it.
I think the phase space was introduced and used a lot by Gibbs before Ehrenfests' article.
Probability is dimensionless number, so if it is to be given by
$$
\int\,\rho\,dq_1dp_1...dq_{3N}dp_{3N}
$$
the probability distribution ##\rho## has to have dimension ##(kgm^2s^{-1})^{-3N}##.
So to define the entropy, which is related to the logarithm of the probabilities (a la Shannon and Jaynes) you need to introduce an arbitrary phase-space measure. In quantum theory one introduces Planck's constant (in modern physics usually the modified Planck constant ##\hbar=h/(2 \pi)##.
Why do we need to introduce arbitrary phase-space measure? What is wrong with the formula
$$
I[\rho] = \int -\rho\ln \rho\, dq_1dp_1...dq_{3N}dp_{3N}
$$
? Logarithm is dimensionless regardless of units in which ##q##'s,##p##'s are measured and ##I## is dimensionless. The only effect units have on ##I## is arbitrary shift of value and additive constant makes no difference in use of ##I##.
but if you calculate the entropy using the Boltzmann statistics under the assumption of distinguishable identical particles, the entropy gets bigger when the diaphragm is taken away, because now it is possible for the particles to change from one half of the box to the other, which they couldn't before. This is of course wrong, because entropy by definition is a state variable and must not depend on the history of the system which brought it finally into the equilibrium situation discussed.
You are comparing two different entropies. It is unfortunate we call many different things with the same name.
The first entropy ##A## is defined as logarithm of volume of phase space compatible with the macrostate. This indeed gets bigger as the constraint is removed and there is no reason why this should not happen in classical physics. Quantity ##A## is not additive; ##A(kU,kV,kN)## is not necessarily equal to ##kA(U,V,N)##.
The second entropy ##B## is thermodynamic entropy defined by integral of ##dQ/T##, which by convention (thus, it does not follow from any physical law) is assumed to be additive. That is, by convention ##B(kU,kV,kN) = kB(U,V,N)##. This assumption is very useful in practice.
A connection between these two entropies can be made. An additive quantity ##A'## can be defined based on ##A## and ##N##-dependent divisor, which gives the same function of macroscopic variables as ##B## is.
$$
B = A' = A/f(N)
$$
Presence of the factor ##f(N)## in this formula is necessitated by the convention imposed on ##B##. This is macroscopic convention. It has no consequences for distinguishability of microscopic particles in principle.