fond
Model Checking Contest @ Petri Nets 2015
Bruxelles, Belgium, June 23, 2015
ITS-Tools compared to other tools («Surprise» models, ReachabilityCardinality)
Last Updated
August 19, 2015

Introduction

This page presents how ITS-Tools do cope efficiently with the ReachabilityCardinality examination face to the other participating tools. In this page, we consider «Surprise» models.

The next sections will show chart comparing performances in termsof both memory and execution time.The x-axis corresponds to the challenging tool where the y-axes represents ITS-Tools' performances. Thus, points below the diagonal of a chart denote comparisons favorables to the tool whileothers corresponds to situations where the challenging tool performs better.

You might also find plots out of the range that denote the case were at least one tool could not answer appropriately (error, time-out, could not compute or did not competed).

ITS-Tools versus GreatSPN-Meddly

Some statistics are displayed below, based on 242 runs (121 for ITS-Tools and 121 for GreatSPN-Meddly, so there are 121 plots on each of the two charts). Each execution was allowed 1 hour and 16 GByte of memory. Then performance charts comparing ITS-Tools to GreatSPN-Meddly are shown (you may click on one graph to enlarge it).

Statistics on the execution
  ITS-Tools GreatSPN-Meddly Both tools   ITS-Tools GreatSPN-Meddly
Computed OK 42 29 3   Smallest Memory Footprint
Do not compete 0 0 0 Times tool wins 42 32
Error detected 55 3 3   Shortest Execution Time
Cannot Compute + Time-out 8 73 10 Times tool wins 43 31


On the chart below, denote cases where the two tools did computed a result, denote the cases where at least one tool did not competed, denote the cases where at least one tool did a mistake and denote the cases where at least one tool stated it could not compute a result or timed-out.

memory chart time chart

ITS-Tools versus LoLA2.0

Some statistics are displayed below, based on 242 runs (121 for ITS-Tools and 121 for LoLA2.0, so there are 121 plots on each of the two charts). Each execution was allowed 1 hour and 16 GByte of memory. Then performance charts comparing ITS-Tools to LoLA2.0 are shown (you may click on one graph to enlarge it).

Statistics on the execution
  ITS-Tools LoLA2.0 Both tools   ITS-Tools LoLA2.0
Computed OK 30 66 15   Smallest Memory Footprint
Do not compete 0 27 0 Times tool wins 31 80
Error detected 53 8 5   Shortest Execution Time
Cannot Compute + Time-out 18 0 0 Times tool wins 37 74


On the chart below, denote cases where the two tools did computed a result, denote the cases where at least one tool did not competed, denote the cases where at least one tool did a mistake and denote the cases where at least one tool stated it could not compute a result or timed-out.

memory chart time chart

ITS-Tools versus LTSMin

Some statistics are displayed below, based on 242 runs (121 for ITS-Tools and 121 for LTSMin, so there are 121 plots on each of the two charts). Each execution was allowed 1 hour and 16 GByte of memory. Then performance charts comparing ITS-Tools to LTSMin are shown (you may click on one graph to enlarge it).

Statistics on the execution
  ITS-Tools LTSMin Both tools   ITS-Tools LTSMin
Computed OK 30 17 15   Smallest Memory Footprint
Do not compete 0 0 0 Times tool wins 33 29
Error detected 13 17 45   Shortest Execution Time
Cannot Compute + Time-out 16 25 2 Times tool wins 40 22


On the chart below, denote cases where the two tools did computed a result, denote the cases where at least one tool did not competed, denote the cases where at least one tool did a mistake and denote the cases where at least one tool stated it could not compute a result or timed-out.

memory chart time chart

ITS-Tools versus Marcie

Some statistics are displayed below, based on 242 runs (121 for ITS-Tools and 121 for Marcie, so there are 121 plots on each of the two charts). Each execution was allowed 1 hour and 16 GByte of memory. Then performance charts comparing ITS-Tools to Marcie are shown (you may click on one graph to enlarge it).

Statistics on the execution
  ITS-Tools Marcie Both tools   ITS-Tools Marcie
Computed OK 34 41 11   Smallest Memory Footprint
Do not compete 0 0 0 Times tool wins 38 48
Error detected 55 3 3   Shortest Execution Time
Cannot Compute + Time-out 8 53 10 Times tool wins 39 47


On the chart below, denote cases where the two tools did computed a result, denote the cases where at least one tool did not competed, denote the cases where at least one tool did a mistake and denote the cases where at least one tool stated it could not compute a result or timed-out.

memory chart time chart

ITS-Tools versus TAPAAL(MC)

Some statistics are displayed below, based on 242 runs (121 for ITS-Tools and 121 for TAPAAL(MC), so there are 121 plots on each of the two charts). Each execution was allowed 1 hour and 16 GByte of memory. Then performance charts comparing ITS-Tools to TAPAAL(MC) are shown (you may click on one graph to enlarge it).

Statistics on the execution
  ITS-Tools TAPAAL(MC) Both tools   ITS-Tools TAPAAL(MC)
Computed OK 24 68 21   Smallest Memory Footprint
Do not compete 0 27 0 Times tool wins 33 80
Error detected 57 1 1   Shortest Execution Time
Cannot Compute + Time-out 17 2 1 Times tool wins 38 75


On the chart below, denote cases where the two tools did computed a result, denote the cases where at least one tool did not competed, denote the cases where at least one tool did a mistake and denote the cases where at least one tool stated it could not compute a result or timed-out.

memory chart time chart

ITS-Tools versus TAPAAL(SEQ)

Some statistics are displayed below, based on 242 runs (121 for ITS-Tools and 121 for TAPAAL(SEQ), so there are 121 plots on each of the two charts). Each execution was allowed 1 hour and 16 GByte of memory. Then performance charts comparing ITS-Tools to TAPAAL(SEQ) are shown (you may click on one graph to enlarge it).

Statistics on the execution
  ITS-Tools TAPAAL(SEQ) Both tools   ITS-Tools TAPAAL(SEQ)
Computed OK 22 71 23   Smallest Memory Footprint
Do not compete 0 27 0 Times tool wins 28 88
Error detected 58 0 0   Shortest Execution Time
Cannot Compute + Time-out 18 0 0 Times tool wins 35 81


On the chart below, denote cases where the two tools did computed a result, denote the cases where at least one tool did not competed, denote the cases where at least one tool did a mistake and denote the cases where at least one tool stated it could not compute a result or timed-out.

memory chart time chart

ITS-Tools versus TAPAAL-OTF(PAR)

Some statistics are displayed below, based on 242 runs (121 for ITS-Tools and 121 for TAPAAL-OTF(PAR), so there are 121 plots on each of the two charts). Each execution was allowed 1 hour and 16 GByte of memory. Then performance charts comparing ITS-Tools to TAPAAL-OTF(PAR) are shown (you may click on one graph to enlarge it).

Statistics on the execution
  ITS-Tools TAPAAL-OTF(PAR) Both tools   ITS-Tools TAPAAL-OTF(PAR)
Computed OK 44 29 1   Smallest Memory Footprint
Do not compete 0 27 0 Times tool wins 44 30
Error detected 55 5 3   Shortest Execution Time
Cannot Compute + Time-out 5 43 13 Times tool wins 44 30


On the chart below, denote cases where the two tools did computed a result, denote the cases where at least one tool did not competed, denote the cases where at least one tool did a mistake and denote the cases where at least one tool stated it could not compute a result or timed-out.

memory chart time chart

ITS-Tools versus TAPAAL-OTF(SEQ)

Some statistics are displayed below, based on 242 runs (121 for ITS-Tools and 121 for TAPAAL-OTF(SEQ), so there are 121 plots on each of the two charts). Each execution was allowed 1 hour and 16 GByte of memory. Then performance charts comparing ITS-Tools to TAPAAL-OTF(SEQ) are shown (you may click on one graph to enlarge it).

Statistics on the execution
  ITS-Tools TAPAAL-OTF(SEQ) Both tools   ITS-Tools TAPAAL-OTF(SEQ)
Computed OK 37 45 8   Smallest Memory Footprint
Do not compete 0 27 0 Times tool wins 43 47
Error detected 58 0 0   Shortest Execution Time
Cannot Compute + Time-out 12 35 6 Times tool wins 39 51


On the chart below, denote cases where the two tools did computed a result, denote the cases where at least one tool did not competed, denote the cases where at least one tool did a mistake and denote the cases where at least one tool stated it could not compute a result or timed-out.

memory chart time chart