Nonlinear experiments involve response and regressors that are connected through a nonlinear regression-type Structure. Examples of nonlinear models include standard nonlinear regression, logistic regression, probit regression. Poisson regression, gamma regression, inverse Gaussian regression, and so on. The Fisher information associated with a nonlinear experiment is typically a complex nonlinear function of the unknown parameter of interest. As a result, we face an awkward situation. Designing an efficient experiment will require knowledge of the parameter, but the purpose of the experiment is to generate data to yield parameter estimates! Our principal objective here is to investigate proper designing of nonlinear experiments that will let us construct efficient estimates of parameters. We focus our attention on a very general nonlinear setup that includes many models commonly encountered in practice. The experiments considered have two fundamental stages: a static design in the initial stage, followed by a fully adaptive sequential stage in which the design points are chosen sequentially, exploiting a D-optimality criterion and using parameter estimates based on available data. We explore the behavior of the maximum likelihood estimate when observations are generated from such an experiment. Two major technical hurdles are ( 1) the dependent nature of the data obtained from an adaptive sequential experiment and (2) the randomness in the total Fisher information associated with the experiment. Our analysis exploits a martingale structure rooted in the likelihood. We derive sufficient conditions that will ensure convergence of the chosen design to a D-optimal one as the number of trials grows. Besides ensuring the large sample optimality of the design, the convergence of the average Fisher information provides an ergodicity condition related to the growth of the martingale processes intrinsically associated with the likelihood. This key observation eventually yields the first-order efficiency of the maximum likelihood estimate via martingale central limit theorem and confirms the asymptotic validity of statistical inference based on the likelihood.