We consider likelihood-based asymptotic inference for a p-dimensional param
eter theta of an identifiable parametric model with singular information ma
trix of rank p - 1 at theta=theta* and likelihood differentiable up to a sp
ecific order. We derive the asymptotic distribution of the likelihood ratio
test statistics for the simple null hypothesis that theta = theta* and of
the maximum likelihood estimator (MLE) of theta when theta = theta*. We sho
w that there exists a reparametrization such that the MLE of the last p - 1
components of theta converges at rate O-p(n(-1/2)). For the first componen
t theta(1) of theta the rate of convergence depends on the order s of the f
irst non-zero partial derivative of the log-likelihood with respect to thet
a(1) evaluated at theta*, When s is odd the rate of convergence of the MLE
of theta(1) is O-p(n(-1/2s)). When s is even, the rate of convergence of th
e MLE of \theta(1)- theta(1)*\ is O-p(n(-1/2s)) and moreover, the asymptoti
c distribution of the sign of the MLE of theta(1) - theta(1)* is non-standa
rd. When p = 1 it is determined by the sign of the sum of the residuals fro
m the population least-squares regression of the (s + l)th derivative of th
e individual contributions to the log-likelihood on their derivatives of or
der s. For p>1, it is determined by a linear combination of the sum of resi
duals of a multivariate population least-squares regression involving parti
al and mixed derivatives of the log-likelihood of a specific order. Thus al
though the MLE of \theta(1) - theta(1)*\ has a uniform rate of convergence
of O-p(n(-1/2s)), the uniform convergence rate for the MLE of theta(1) in s
uitable shrinking neighbourhoods of theta(1)* is only O-p(n(-1/(2s+2))).