fond
Model Checking Contest 2019
9th edition, Prague, Czech Republic, April 7, 2019 (TOOLympics)
Execution of r198-smll-155272319100060
Last Updated
Apr 15, 2019

About the Execution of ITS-Tools.M for PhilosophersDyn-PT-10

Execution Summary
Max Memory
Used (MB)
Time wait (ms) CPU Usage (ms) I/O Wait (ms) Computed Result Execution
Status
5336.410 343336.00 1270207.00 353.00 FFTFFTTFTFFFFTTT normal

Execution Chart

We display below the execution chart for this examination (boot time has been removed).

Trace from the execution

Formatting '/data/fkordon/mcc2019-input.r198-smll-155272319100060.qcow2', fmt=qcow2 size=4294967296 backing_file='/data/fkordon/mcc2019-input.qcow2' encryption=off cluster_size=65536 lazy_refcounts=off
Waiting for the VM to be ready (probing ssh)
......................
=====================================================================
Generated by BenchKit 2-3957
Executing tool itstoolsm
Input is PhilosophersDyn-PT-10, examination is LTLCardinality
Time confinement is 3600 seconds
Memory confinement is 16384 MBytes
Number of cores is 4
Run identifier is r198-smll-155272319100060
=====================================================================

--------------------
preparation of the directory to be used:
/home/mcc/execution
total 4.7M
-rw-r--r-- 1 mcc users 20K Feb 12 09:13 CTLCardinality.txt
-rw-r--r-- 1 mcc users 59K Feb 12 09:13 CTLCardinality.xml
-rw-r--r-- 1 mcc users 81K Feb 8 08:59 CTLFireability.txt
-rw-r--r-- 1 mcc users 303K Feb 8 08:59 CTLFireability.xml
-rw-r--r-- 1 mcc users 4.0K Mar 10 17:31 GenericPropertiesDefinition.xml
-rw-r--r-- 1 mcc users 5.9K Mar 10 17:31 GenericPropertiesVerdict.xml
-rw-r--r-- 1 mcc users 108 Feb 24 15:05 GlobalProperties.txt
-rw-r--r-- 1 mcc users 346 Feb 24 15:05 GlobalProperties.xml
-rw-r--r-- 1 mcc users 12K Feb 5 00:32 LTLCardinality.txt
-rw-r--r-- 1 mcc users 38K Feb 5 00:32 LTLCardinality.xml
-rw-r--r-- 1 mcc users 82K Feb 4 22:38 LTLFireability.txt
-rw-r--r-- 1 mcc users 295K Feb 4 22:38 LTLFireability.xml
-rw-r--r-- 1 mcc users 22K Feb 4 12:40 ReachabilityCardinality.txt
-rw-r--r-- 1 mcc users 66K Feb 4 12:40 ReachabilityCardinality.xml
-rw-r--r-- 1 mcc users 251K Feb 1 07:43 ReachabilityFireability.txt
-rw-r--r-- 1 mcc users 952K Feb 1 07:43 ReachabilityFireability.xml
-rw-r--r-- 1 mcc users 2.7K Feb 4 22:23 UpperBounds.txt
-rw-r--r-- 1 mcc users 6.1K Feb 4 22:23 UpperBounds.xml

-rw-r--r-- 1 mcc users 5 Jan 29 09:34 equiv_col
-rw-r--r-- 1 mcc users 3 Jan 29 09:34 instance
-rw-r--r-- 1 mcc users 6 Jan 29 09:34 iscolored
-rw-r--r-- 1 mcc users 2.5M Mar 10 17:31 model.pnml

--------------------
content from stdout:

=== Data for post analysis generated by BenchKit (invocation template)

The expected result is a vector of booleans
BOOL_VECTOR

here is the order used to build the result vector(from text file)
FORMULA_NAME PhilosophersDyn-PT-10-LTLCardinality-00
FORMULA_NAME PhilosophersDyn-PT-10-LTLCardinality-01
FORMULA_NAME PhilosophersDyn-PT-10-LTLCardinality-02
FORMULA_NAME PhilosophersDyn-PT-10-LTLCardinality-03
FORMULA_NAME PhilosophersDyn-PT-10-LTLCardinality-04
FORMULA_NAME PhilosophersDyn-PT-10-LTLCardinality-05
FORMULA_NAME PhilosophersDyn-PT-10-LTLCardinality-06
FORMULA_NAME PhilosophersDyn-PT-10-LTLCardinality-07
FORMULA_NAME PhilosophersDyn-PT-10-LTLCardinality-08
FORMULA_NAME PhilosophersDyn-PT-10-LTLCardinality-09
FORMULA_NAME PhilosophersDyn-PT-10-LTLCardinality-10
FORMULA_NAME PhilosophersDyn-PT-10-LTLCardinality-11
FORMULA_NAME PhilosophersDyn-PT-10-LTLCardinality-12
FORMULA_NAME PhilosophersDyn-PT-10-LTLCardinality-13
FORMULA_NAME PhilosophersDyn-PT-10-LTLCardinality-14
FORMULA_NAME PhilosophersDyn-PT-10-LTLCardinality-15

=== Now, execution of the tool begins

BK_START 1553552936408

Working with output stream class java.io.PrintStream
Using solver Z3 to compute partial order matrices.
Built C files in :
/home/mcc/execution
Running greatSPN : CommandLine [args=[/home/mcc/BenchKit//greatspn//bin/pinvar, /home/mcc/execution/gspn], workingDir=/home/mcc/execution]
Run of greatSPN captured in /home/mcc/execution/outPut.txt
Running greatSPN : CommandLine [args=[/home/mcc/BenchKit//greatspn//bin/RGMEDD2, /home/mcc/execution/gspn, -META, -varord-only], workingDir=/home/mcc/execution]
Run of greatSPN captured in /home/mcc/execution/outPut.txt
Running compilation step : CommandLine [args=[gcc, -c, -I/home/mcc/BenchKit//lts_install_dir//include, -I., -std=c99, -fPIC, -O2, model.c], workingDir=/home/mcc/execution]
Using order generated by GreatSPN with heuristic : META
Invoking ITS tools like this :CommandLine [args=[/home/mcc/BenchKit/itstools/plugins/fr.lip6.move.gal.itstools.binaries_1.0.0.201903251645/bin/its-ltl-linux64, --gc-threshold, 2000000, -i, /home/mcc/execution/LTLCardinality.pnml.gal, -t, CGAL, -LTL, /home/mcc/execution/LTLCardinality.ltl, -c, -stutter-deadlock, --load-order, /home/mcc/execution/model.ord], workingDir=/home/mcc/execution]

its-ltl command run as :

/home/mcc/BenchKit/itstools/plugins/fr.lip6.move.gal.itstools.binaries_1.0.0.201903251645/bin/its-ltl-linux64 --gc-threshold 2000000 -i /home/mcc/execution/LTLCardinality.pnml.gal -t CGAL -LTL /home/mcc/execution/LTLCardinality.ltl -c -stutter-deadlock --load-order /home/mcc/execution/model.ord
Read 16 LTL properties
Successfully loaded order from file /home/mcc/execution/model.ord
Checking formula 0 : !((G(("((((((((((Outside_10+Outside_4)+Outside_3)+Outside_5)+Outside_6)+Outside_7)+Outside_8)+Outside_9)+Outside_2)+Outside_1)<=(((((((((WaitLeft_7+WaitLeft_4)+WaitLeft_10)+WaitLeft_5)+WaitLeft_2)+WaitLeft_3)+WaitLeft_1)+WaitLeft_8)+WaitLeft_9)+WaitLeft_6))")U(G("((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((Neighbourhood_7_3+Neighbourhood_2_10)+Neighbourhood_9_1)+Neighbourhood_3_8)+Neighbourhood_8_2)+Neighbourhood_5_5)+Neighbourhood_8_9)+Neighbourhood_3_4)+Neighbourhood_7_7)+Neighbourhood_3_6)+Neighbourhood_4_8)+Neighbourhood_5_10)+Neighbourhood_5_1)+Neighbourhood_6_5)+Neighbourhood_1_9)+Neighbourhood_1_7)+Neighbourhood_6_3)+Neighbourhood_9_2)+Neighbourhood_2_2)+Neighbourhood_10_4)+Neighbourhood_7_5)+Neighbourhood_1_2)+Neighbourhood_5_8)+Neighbourhood_6_1)+Neighbourhood_9_9)+Neighbourhood_2_7)+Neighbourhood_6_7)+Neighbourhood_10_2)+Neighbourhood_4_4)+Neighbourhood_2_6)+Neighbourhood_5_3)+Neighbourhood_4_10)+Neighbourhood_3_1)+Neighbourhood_9_4)+Neighbourhood_9_7)+Neighbourhood_8_5)+Neighbourhood_4_3)+Neighbourhood_6_8)+Neighbourhood_9_6)+Neighbourhood_10_8)+Neighbourhood_1_6)+Neighbourhood_9_8)+Neighbourhood_1_4)+Neighbourhood_6_9)+Neighbourhood_7_10)+Neighbourhood_8_7)+Neighbourhood_7_8)+Neighbourhood_4_1)+Neighbourhood_6_10)+Neighbourhood_4_2)+Neighbourhood_3_3)+Neighbourhood_1_5)+Neighbourhood_2_4)+Neighbourhood_8_8)+Neighbourhood_7_9)+Neighbourhood_10_6)+Neighbourhood_10_5)+Neighbourhood_3_2)+Neighbourhood_9_3)+Neighbourhood_4_7)+Neighbourhood_6_4)+Neighbourhood_2_3)+Neighbourhood_5_2)+Neighbourhood_6_6)+Neighbourhood_8_1)+Neighbourhood_10_3)+Neighbourhood_2_5)+Neighbourhood_1_1)+Neighbourhood_10_7)+Neighbourhood_1_3)+Neighbourhood_9_5)+Neighbourhood_7_1)+Neighbourhood_3_7)+Neighbourhood_8_6)+Neighbourhood_6_2)+Neighbourhood_5_4)+Neighbourhood_8_10)+Neighbourhood_5_9)+Neighbourhood_9_10)+Neighbourhood_4_9)+Neighbourhood_7_6)+Neighbourhood_2_1)+Neighbourhood_1_8)+Neighbourhood_3_5)+Neighbourhood_10_9)+Neighbourhood_3_9)+Neighbourhood_8_4)+Neighbourhood_5_6)+Neighbourhood_10_1)+Neighbourhood_7_2)+Neighbourhood_4_5)+Neighbourhood_2_8)+Neighbourhood_10_10)+Neighbourhood_4_6)+Neighbourhood_7_4)+Neighbourhood_8_3)+Neighbourhood_1_10)+Neighbourhood_2_9)+Neighbourhood_5_7)+Neighbourhood_3_10)<=(((((((((WaitRight_5+WaitRight_10)+WaitRight_1)+WaitRight_9)+WaitRight_2)+WaitRight_8)+WaitRight_3)+WaitRight_7)+WaitRight_6)+WaitRight_4))")))))
Formula 0 simplified : !G("((((((((((Outside_10+Outside_4)+Outside_3)+Outside_5)+Outside_6)+Outside_7)+Outside_8)+Outside_9)+Outside_2)+Outside_1)<=(((((((((WaitLeft_7+WaitLeft_4)+WaitLeft_10)+WaitLeft_5)+WaitLeft_2)+WaitLeft_3)+WaitLeft_1)+WaitLeft_8)+WaitLeft_9)+WaitLeft_6))" U G"((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((Neighbourhood_7_3+Neighbourhood_2_10)+Neighbourhood_9_1)+Neighbourhood_3_8)+Neighbourhood_8_2)+Neighbourhood_5_5)+Neighbourhood_8_9)+Neighbourhood_3_4)+Neighbourhood_7_7)+Neighbourhood_3_6)+Neighbourhood_4_8)+Neighbourhood_5_10)+Neighbourhood_5_1)+Neighbourhood_6_5)+Neighbourhood_1_9)+Neighbourhood_1_7)+Neighbourhood_6_3)+Neighbourhood_9_2)+Neighbourhood_2_2)+Neighbourhood_10_4)+Neighbourhood_7_5)+Neighbourhood_1_2)+Neighbourhood_5_8)+Neighbourhood_6_1)+Neighbourhood_9_9)+Neighbourhood_2_7)+Neighbourhood_6_7)+Neighbourhood_10_2)+Neighbourhood_4_4)+Neighbourhood_2_6)+Neighbourhood_5_3)+Neighbourhood_4_10)+Neighbourhood_3_1)+Neighbourhood_9_4)+Neighbourhood_9_7)+Neighbourhood_8_5)+Neighbourhood_4_3)+Neighbourhood_6_8)+Neighbourhood_9_6)+Neighbourhood_10_8)+Neighbourhood_1_6)+Neighbourhood_9_8)+Neighbourhood_1_4)+Neighbourhood_6_9)+Neighbourhood_7_10)+Neighbourhood_8_7)+Neighbourhood_7_8)+Neighbourhood_4_1)+Neighbourhood_6_10)+Neighbourhood_4_2)+Neighbourhood_3_3)+Neighbourhood_1_5)+Neighbourhood_2_4)+Neighbourhood_8_8)+Neighbourhood_7_9)+Neighbourhood_10_6)+Neighbourhood_10_5)+Neighbourhood_3_2)+Neighbourhood_9_3)+Neighbourhood_4_7)+Neighbourhood_6_4)+Neighbourhood_2_3)+Neighbourhood_5_2)+Neighbourhood_6_6)+Neighbourhood_8_1)+Neighbourhood_10_3)+Neighbourhood_2_5)+Neighbourhood_1_1)+Neighbourhood_10_7)+Neighbourhood_1_3)+Neighbourhood_9_5)+Neighbourhood_7_1)+Neighbourhood_3_7)+Neighbourhood_8_6)+Neighbourhood_6_2)+Neighbourhood_5_4)+Neighbourhood_8_10)+Neighbourhood_5_9)+Neighbourhood_9_10)+Neighbourhood_4_9)+Neighbourhood_7_6)+Neighbourhood_2_1)+Neighbourhood_1_8)+Neighbourhood_3_5)+Neighbourhood_10_9)+Neighbourhood_3_9)+Neighbourhood_8_4)+Neighbourhood_5_6)+Neighbourhood_10_1)+Neighbourhood_7_2)+Neighbourhood_4_5)+Neighbourhood_2_8)+Neighbourhood_10_10)+Neighbourhood_4_6)+Neighbourhood_7_4)+Neighbourhood_8_3)+Neighbourhood_1_10)+Neighbourhood_2_9)+Neighbourhood_5_7)+Neighbourhood_3_10)<=(((((((((WaitRight_5+WaitRight_10)+WaitRight_1)+WaitRight_9)+WaitRight_2)+WaitRight_8)+WaitRight_3)+WaitRight_7)+WaitRight_6)+WaitRight_4))")
Compilation finished in 38082 ms.
Running link step : CommandLine [args=[gcc, -shared, -o, gal.so, model.o], workingDir=/home/mcc/execution]
Link finished in 77 ms.
Running LTSmin : CommandLine [args=[/home/mcc/BenchKit//lts_install_dir//bin/pins2lts-mc, ./gal.so, --threads=8, -p, --pins-guards, --when, --ltl, [](((LTLAP0==true))U([]((LTLAP1==true)))), --buchi-type=spotba], workingDir=/home/mcc/execution]
LTSmin run took 38392 ms.
FORMULA PhilosophersDyn-PT-10-LTLCardinality-00 FALSE TECHNIQUES PARTIAL_ORDER EXPLICIT LTSMIN SAT_SMT
Running LTSmin : CommandLine [args=[/home/mcc/BenchKit//lts_install_dir//bin/pins2lts-mc, ./gal.so, --threads=8, --when, --ltl, X((X((LTLAP2==true)))U(X((LTLAP3==true)))), --buchi-type=spotba], workingDir=/home/mcc/execution]
LTSmin run took 558 ms.
FORMULA PhilosophersDyn-PT-10-LTLCardinality-01 FALSE TECHNIQUES PARTIAL_ORDER EXPLICIT LTSMIN SAT_SMT
Running LTSmin : CommandLine [args=[/home/mcc/BenchKit//lts_install_dir//bin/pins2lts-mc, ./gal.so, --threads=8, -p, --pins-guards, --when, --ltl, (<>(<>((LTLAP4==true))))U(<>((LTLAP5==true))), --buchi-type=spotba], workingDir=/home/mcc/execution]
LTSmin run took 36873 ms.
FORMULA PhilosophersDyn-PT-10-LTLCardinality-02 TRUE TECHNIQUES PARTIAL_ORDER EXPLICIT LTSMIN SAT_SMT
Running LTSmin : CommandLine [args=[/home/mcc/BenchKit//lts_install_dir//bin/pins2lts-mc, ./gal.so, --threads=8, --when, --ltl, [](([]((LTLAP6==true)))U(X((LTLAP7==true)))), --buchi-type=spotba], workingDir=/home/mcc/execution]
LTSmin run took 263 ms.
FORMULA PhilosophersDyn-PT-10-LTLCardinality-03 FALSE TECHNIQUES PARTIAL_ORDER EXPLICIT LTSMIN SAT_SMT
Running LTSmin : CommandLine [args=[/home/mcc/BenchKit//lts_install_dir//bin/pins2lts-mc, ./gal.so, --threads=8, --when, --ltl, [](X(((LTLAP8==true))U((LTLAP7==true)))), --buchi-type=spotba], workingDir=/home/mcc/execution]
LTSmin run took 208 ms.
FORMULA PhilosophersDyn-PT-10-LTLCardinality-04 FALSE TECHNIQUES PARTIAL_ORDER EXPLICIT LTSMIN SAT_SMT
Running LTSmin : CommandLine [args=[/home/mcc/BenchKit//lts_install_dir//bin/pins2lts-mc, ./gal.so, --threads=8, --when, --ltl, X(<>(((LTLAP9==true))U((LTLAP10==true)))), --buchi-type=spotba], workingDir=/home/mcc/execution]
LTSmin run took 484 ms.
FORMULA PhilosophersDyn-PT-10-LTLCardinality-05 TRUE TECHNIQUES PARTIAL_ORDER EXPLICIT LTSMIN SAT_SMT
Running LTSmin : CommandLine [args=[/home/mcc/BenchKit//lts_install_dir//bin/pins2lts-mc, ./gal.so, --threads=8, --when, --ltl, <>(((LTLAP11==true))U(X((LTLAP12==true)))), --buchi-type=spotba], workingDir=/home/mcc/execution]
LTSmin run took 515 ms.
FORMULA PhilosophersDyn-PT-10-LTLCardinality-06 TRUE TECHNIQUES PARTIAL_ORDER EXPLICIT LTSMIN SAT_SMT
Running LTSmin : CommandLine [args=[/home/mcc/BenchKit//lts_install_dir//bin/pins2lts-mc, ./gal.so, --threads=8, --when, --ltl, X(<>(<>([]((LTLAP13==true))))), --buchi-type=spotba], workingDir=/home/mcc/execution]
LTSmin run took 484 ms.
FORMULA PhilosophersDyn-PT-10-LTLCardinality-07 FALSE TECHNIQUES PARTIAL_ORDER EXPLICIT LTSMIN SAT_SMT
Running LTSmin : CommandLine [args=[/home/mcc/BenchKit//lts_install_dir//bin/pins2lts-mc, ./gal.so, --threads=8, -p, --pins-guards, --when, --ltl, <>((<>((LTLAP14==true)))U(<>((LTLAP15==true)))), --buchi-type=spotba], workingDir=/home/mcc/execution]
LTSmin run took 36400 ms.
FORMULA PhilosophersDyn-PT-10-LTLCardinality-08 TRUE TECHNIQUES PARTIAL_ORDER EXPLICIT LTSMIN SAT_SMT
Running LTSmin : CommandLine [args=[/home/mcc/BenchKit//lts_install_dir//bin/pins2lts-mc, ./gal.so, --threads=8, --when, --ltl, (((LTLAP16==true))U((LTLAP17==true)))U(X([]((LTLAP18==true)))), --buchi-type=spotba], workingDir=/home/mcc/execution]
LTSmin run took 705 ms.
FORMULA PhilosophersDyn-PT-10-LTLCardinality-09 FALSE TECHNIQUES PARTIAL_ORDER EXPLICIT LTSMIN SAT_SMT
Running LTSmin : CommandLine [args=[/home/mcc/BenchKit//lts_install_dir//bin/pins2lts-mc, ./gal.so, --threads=8, -p, --pins-guards, --when, --ltl, <>([](((LTLAP19==true))U((LTLAP20==true)))), --buchi-type=spotba], workingDir=/home/mcc/execution]
LTSmin run took 36155 ms.
FORMULA PhilosophersDyn-PT-10-LTLCardinality-10 FALSE TECHNIQUES PARTIAL_ORDER EXPLICIT LTSMIN SAT_SMT
Running LTSmin : CommandLine [args=[/home/mcc/BenchKit//lts_install_dir//bin/pins2lts-mc, ./gal.so, --threads=8, -p, --pins-guards, --when, --ltl, []((LTLAP21==true)), --buchi-type=spotba], workingDir=/home/mcc/execution]
LTSmin run took 38476 ms.
FORMULA PhilosophersDyn-PT-10-LTLCardinality-11 FALSE TECHNIQUES PARTIAL_ORDER EXPLICIT LTSMIN SAT_SMT
Running LTSmin : CommandLine [args=[/home/mcc/BenchKit//lts_install_dir//bin/pins2lts-mc, ./gal.so, --threads=8, --when, --ltl, <>(X(((LTLAP22==true))U((LTLAP23==true)))), --buchi-type=spotba], workingDir=/home/mcc/execution]
LTSmin run took 367 ms.
FORMULA PhilosophersDyn-PT-10-LTLCardinality-12 FALSE TECHNIQUES PARTIAL_ORDER EXPLICIT LTSMIN SAT_SMT
Running LTSmin : CommandLine [args=[/home/mcc/BenchKit//lts_install_dir//bin/pins2lts-mc, ./gal.so, --threads=8, -p, --pins-guards, --when, --ltl, true, --buchi-type=spotba], workingDir=/home/mcc/execution]
LTSmin run took 35838 ms.
FORMULA PhilosophersDyn-PT-10-LTLCardinality-13 TRUE TECHNIQUES PARTIAL_ORDER EXPLICIT LTSMIN SAT_SMT
Running LTSmin : CommandLine [args=[/home/mcc/BenchKit//lts_install_dir//bin/pins2lts-mc, ./gal.so, --threads=8, -p, --pins-guards, --when, --ltl, <>((LTLAP24==true)), --buchi-type=spotba], workingDir=/home/mcc/execution]
LTSmin run took 35632 ms.
FORMULA PhilosophersDyn-PT-10-LTLCardinality-14 TRUE TECHNIQUES PARTIAL_ORDER EXPLICIT LTSMIN SAT_SMT
Running LTSmin : CommandLine [args=[/home/mcc/BenchKit//lts_install_dir//bin/pins2lts-mc, ./gal.so, --threads=8, -p, --pins-guards, --when, --ltl, true, --buchi-type=spotba], workingDir=/home/mcc/execution]
LTSmin run took 35943 ms.
FORMULA PhilosophersDyn-PT-10-LTLCardinality-15 TRUE TECHNIQUES PARTIAL_ORDER EXPLICIT LTSMIN SAT_SMT
ITS tools runner thread asked to quit. Dying gracefully.

BK_STOP 1553553279744

--------------------
content from stderr:

+ export BINDIR=/home/mcc/BenchKit/
+ BINDIR=/home/mcc/BenchKit/
++ pwd
+ export MODEL=/home/mcc/execution
+ MODEL=/home/mcc/execution
+ [[ LTLCardinality = StateSpace ]]
+ /home/mcc/BenchKit//runeclipse.sh /home/mcc/execution LTLCardinality -its -ltsminpath /home/mcc/BenchKit//lts_install_dir/ -greatspnpath /home/mcc/BenchKit//greatspn/ -order META -manyOrder -smt
+ ulimit -s 65536
+ [[ -z '' ]]
+ export LTSMIN_MEM_SIZE=8589934592
+ LTSMIN_MEM_SIZE=8589934592
+ /home/mcc/BenchKit//itstools/its-tools -data /home/mcc/execution/workspace -pnfolder /home/mcc/execution -examination LTLCardinality -z3path /home/mcc/BenchKit//z3/bin/z3 -yices2path /home/mcc/BenchKit//yices/bin/yices -its -ltsminpath /home/mcc/BenchKit//lts_install_dir/ -greatspnpath /home/mcc/BenchKit//greatspn/ -order META -manyOrder -smt -vmargs -Dosgi.locking=none -Declipse.stateSaveDelayInterval=-1 -Dosgi.configuration.area=/tmp/.eclipse -Xss8m -Xms40m -Xmx8192m -Dfile.encoding=UTF-8 -Dosgi.requiredJavaVersion=1.6
Mar 25, 2019 10:28:59 PM fr.lip6.move.gal.application.Application start
INFO: Running its-tools with arguments : [-pnfolder, /home/mcc/execution, -examination, LTLCardinality, -z3path, /home/mcc/BenchKit//z3/bin/z3, -yices2path, /home/mcc/BenchKit//yices/bin/yices, -its, -ltsminpath, /home/mcc/BenchKit//lts_install_dir/, -greatspnpath, /home/mcc/BenchKit//greatspn/, -order, META, -manyOrder, -smt]
Mar 25, 2019 10:28:59 PM fr.lip6.move.gal.application.MccTranslator transformPNML
INFO: Parsing pnml file : /home/mcc/execution/model.pnml
Mar 25, 2019 10:28:59 PM fr.lip6.move.gal.nupn.PTNetReader loadFromXML
INFO: Load time of PNML (sax parser for PT used): 340 ms
Mar 25, 2019 10:28:59 PM fr.lip6.move.gal.pnml.togal.PTGALTransformer handlePage
INFO: Transformed 170 places.
Mar 25, 2019 10:29:00 PM fr.lip6.move.gal.pnml.togal.PTGALTransformer handlePage
INFO: Transformed 2310 transitions.
Mar 25, 2019 10:29:01 PM fr.lip6.move.gal.instantiate.GALRewriter flatten
INFO: Flatten gal took : 1228 ms
Mar 25, 2019 10:29:02 PM fr.lip6.move.gal.instantiate.GALRewriter flatten
INFO: Flatten gal took : 510 ms
Mar 25, 2019 10:29:02 PM fr.lip6.move.gal.semantics.DeterministicNextBuilder getDeterministicNext
INFO: Input system was already deterministic with 2310 transitions.
Mar 25, 2019 10:29:02 PM fr.lip6.move.gal.gal2pins.Gal2PinsTransformerNext transform
INFO: Too many transitions (2310) to apply POR reductions. Disabling POR matrices.
Mar 25, 2019 10:29:02 PM fr.lip6.move.gal.semantics.DeterministicNextBuilder getDeterministicNext
INFO: Input system was already deterministic with 2310 transitions.
Mar 25, 2019 10:29:02 PM fr.lip6.move.gal.application.StructuralToGreatSPN handlePage
INFO: Transformed 170 places.
Mar 25, 2019 10:29:02 PM fr.lip6.move.gal.application.StructuralToGreatSPN handlePage
INFO: Transformed 2310 transitions.
Mar 25, 2019 10:29:03 PM fr.lip6.move.gal.gal2pins.Gal2PinsTransformerNext transform
INFO: Built C files in 1362ms conformant to PINS in folder :/home/mcc/execution
Mar 25, 2019 10:29:03 PM fr.lip6.move.serialization.SerializationUtil systemToFile
INFO: Time to serialize gal into /home/mcc/execution/LTLCardinality.pnml.gal : 35 ms
Mar 25, 2019 10:29:03 PM fr.lip6.move.serialization.SerializationUtil serializePropertiesForITSLTLTools
INFO: Time to serialize properties into /home/mcc/execution/LTLCardinality.ltl : 2 ms

Sequence of Actions to be Executed by the VM

This is useful if one wants to reexecute the tool in the VM from the submitted image disk.

set -x
# this is for BenchKit: configuration of major elements for the test
export BK_INPUT="PhilosophersDyn-PT-10"
export BK_EXAMINATION="LTLCardinality"
export BK_TOOL="itstoolsm"
export BK_RESULT_DIR="/tmp/BK_RESULTS/OUTPUTS"
export BK_TIME_CONFINEMENT="3600"
export BK_MEMORY_CONFINEMENT="16384"

# this is specific to your benchmark or test

export BIN_DIR="$HOME/BenchKit/bin"

# remove the execution directoty if it exists (to avoid increse of .vmdk images)
if [ -d execution ] ; then
rm -rf execution
fi

# this is for BenchKit: explicit launching of the test
echo "====================================================================="
echo " Generated by BenchKit 2-3957"
echo " Executing tool itstoolsm"
echo " Input is PhilosophersDyn-PT-10, examination is LTLCardinality"
echo " Time confinement is $BK_TIME_CONFINEMENT seconds"
echo " Memory confinement is 16384 MBytes"
echo " Number of cores is 4"
echo " Run identifier is r198-smll-155272319100060"
echo "====================================================================="
echo
echo "--------------------"
echo "preparation of the directory to be used:"

tar xzf /home/mcc/BenchKit/INPUTS/PhilosophersDyn-PT-10.tgz
mv PhilosophersDyn-PT-10 execution
cd execution
if [ "LTLCardinality" = "GlobalProperties" ] ; then
rm -f GenericPropertiesVerdict.xml
fi
if [ "LTLCardinality" = "UpperBounds" ] ; then
rm -f GenericPropertiesVerdict.xml
fi
pwd
ls -lh

echo
echo "--------------------"
echo "content from stdout:"
echo
echo "=== Data for post analysis generated by BenchKit (invocation template)"
echo
if [ "LTLCardinality" = "UpperBounds" ] ; then
echo "The expected result is a vector of positive values"
echo NUM_VECTOR
elif [ "LTLCardinality" != "StateSpace" ] ; then
echo "The expected result is a vector of booleans"
echo BOOL_VECTOR
else
echo "no data necessary for post analysis"
fi
echo
if [ -f "LTLCardinality.txt" ] ; then
echo "here is the order used to build the result vector(from text file)"
for x in $(grep Property LTLCardinality.txt | cut -d ' ' -f 2 | sort -u) ; do
echo "FORMULA_NAME $x"
done
elif [ -f "LTLCardinality.xml" ] ; then # for cunf (txt files deleted;-)
echo echo "here is the order used to build the result vector(from xml file)"
for x in $(grep '' LTLCardinality.xml | cut -d '>' -f 2 | cut -d '<' -f 1 | sort -u) ; do
echo "FORMULA_NAME $x"
done
fi
echo
echo "=== Now, execution of the tool begins"
echo
echo -n "BK_START "
date -u +%s%3N
echo
timeout -s 9 $BK_TIME_CONFINEMENT bash -c "/home/mcc/BenchKit/BenchKit_head.sh 2> STDERR ; echo ; echo -n \"BK_STOP \" ; date -u +%s%3N"
if [ $? -eq 137 ] ; then
echo
echo "BK_TIME_CONFINEMENT_REACHED"
fi
echo
echo "--------------------"
echo "content from stderr:"
echo
cat STDERR ;