This paper presents a method for describing the effect of articulatory trajectories on phoneme recognition. The proposed method comprises three stages. The first stage embeds three multilayer neural networks (MLNs): MLN (LF-DPF) that maps acoustic features or local features (LFs) onto articulatory features or distinctive phonetic features (DPFs), MLN (cntxt) that reduces misclassifications at phoneme boundaries, and MLN (Dyn) that controls dynamics of DPF features. The second stage incorporates an inhibition/enhancement (In/En) network by varying the trajectories of articulators to achieve categorical DPF movement by enhancing DPF peak patterns and inhibiting DPF dip patterns. The third stage decor relates continuous DPF vectors using the Gram-Schmidt algorithm before feeding into a hidden Markov model (HMM)-based classifier. In the experiments on Japanese Newspaper Article Sentences (JNAS) database, the proposed feature extractor shows the performance of phoneme recognition with the variation of different trajectories.