Runtime verification is concerned with monitoring and analysis of software and hardware system executions. Runtime verification techniques are crucial for system correctness, reliability, and robustness; they are significantly more powerful and versatile than conventional testing, and more practical than exhaustive formal verification. Runtime verification can be used prior to deployment, for testing, verification, and debugging purposes, and after deployment for ensuring reliability, safety, and security and for providing fault containment and recovery as well as online system repair. Topics of interest to the conference include:

Application areas of runtime verification include safety/mission-critical systems, enterprise and systems software, autonomous and reactive control systems, health management and diagnosis systems, and system security and privacy.

Technical Research Papers Track:

Technical research papers can be submitted in two categories: regular and short papers. Papers in both categories will be reviewed by the conference Program Committee. All accepted technical papers will appear in an LNCS volume. Submitted papers must use the LNCS style. At least one author of each accepted paper must attend RV’14 to present the paper. Papers must be submitted electronically using the EasyChair system.

A blind review process will be enforced. Authors should not reveal authorship directly or indirectly through references. For self-citations, authors can either use third-person verbs and pronouns or use the phrase “citation withheld due to the double-blind process”.

Regular papers (up to 15 pages) should present original unpublished results. Theoretical and experimental papers as well as papers on applications of runtime verification and case studies are all welcome. A non-monetary Best Paper Award will be given. A selection of accepted regular papers will be invited to appear in a special issue of the Springer Journal on Formal Methods in System Design.

Short papers (up to 5 pages) may present novel but not necessarily thoroughly worked out ideas, for example emerging runtime verification techniques and applications, or techniques and applications that establish relationships between runtime verification and other domains. Accepted short papers will be presented in special short talk (10 minutes) and poster sessions.

The aim of the RV 2014 tool demonstration track is to provide an opportunity for researchers and practitioners to show and to discuss the latest advances, experiences and challenges in devising and developing reliable software tools for runtime verification. Tool demonstration papers will be reviewed by the Tools Track Program Committee. All accepted tool demonstration papers will appear in the conference proceedings LNCS volume. Submitted papers must use the LNCS style. At least one author of each accepted paper must attend RV’14 to present the paper.Papers must be submitted electronically using theEasyChairsystem. Submissions in this track are not subject to blind review.

Tool papers should meet the following criteria:

A tool paper should present a new tool, a new tool component or novel extensions to existing tools supporting runtime verification. Each submission should be original and not published previously in a tool paper form.

Each submission must not exceed 8 pages in the LNCS/Springer proceeding format, including all text, references and figures. The paper must be written in English and provided in PDF format.

Each submission must be accompanied at the time of the submission by a short screencast (between 5-10 minutes), with voice and overlay text commentary illustrating the demonstration of the tool (a link to it should be provided in the paper).

The paper must include information on tool availability, maturity, selected experimental results and it should provide a link to a website containing the theoretical background and user guide. Furthermore, we strongly encourage authors to make their tools and benchmarks available with their submission.

Each tool paper must include a script in an appendix (not included in the page count) describing how the demo will be conducted during the conference presentation with screenshots presenting step-by-step the tool’s capabilities, highlighting the main characteristics and the usage.

Evaluation

Each submission will be reviewed by at least four members of the tool demonstration track program committee. The evaluation criteria will include: