We seek to quantify the failure and success of dynamic Bayesian networks (DBNs), a popular tool for reverse-engineering networks from time-series data. In particular, we focus on data generated by continuous time processes (e.g., genetic expression) and sampled at discrete times. To facilitate analysis and interpretation, we employ a "minimal model" to generate arbitrary abundances of stochastic data from networks of known topologies, which are then sub-sampled and in some cases interpolated. We find that DBNs perform relatively poorly when given data sets comparable to those used for genetic network inference. Interpolation does not appear to improve inference success. Finally, we contrast the performance of DBNs with results from linear regression on our synthetic data.