About the Execution of ITS-Tools.L for PhilosophersDyn-PT-10
Execution Summary | |||||
Max Memory Used (MB) |
Time wait (ms) | CPU Usage (ms) | I/O Wait (ms) | Computed Result | Execution Status |
15740.860 | 182840.00 | 374309.00 | 164.90 | FFFFTTTFTTTFFTTF | normal |
Execution Chart
We display below the execution chart for this examination (boot time has been removed).
Trace from the execution
Waiting for the VM to be ready (probing ssh)
...................
/home/mcc/execution
total 4.8M
-rw-r--r-- 1 mcc users 19K May 15 18:54 CTLCardinality.txt
-rw-r--r-- 1 mcc users 63K May 15 18:54 CTLCardinality.xml
-rw-r--r-- 1 mcc users 151K May 15 18:54 CTLFireability.txt
-rw-r--r-- 1 mcc users 585K May 15 18:54 CTLFireability.xml
-rw-r--r-- 1 mcc users 4.0K May 15 18:50 GenericPropertiesDefinition.xml
-rw-r--r-- 1 mcc users 5.9K May 15 18:50 GenericPropertiesVerdict.xml
-rw-r--r-- 1 mcc users 8.9K May 26 09:27 LTLCardinality.txt
-rw-r--r-- 1 mcc users 29K May 26 09:27 LTLCardinality.xml
-rw-r--r-- 1 mcc users 58K May 26 09:27 LTLFireability.txt
-rw-r--r-- 1 mcc users 212K May 26 09:27 LTLFireability.xml
-rw-r--r-- 1 mcc users 21K May 15 18:54 ReachabilityCardinality.txt
-rw-r--r-- 1 mcc users 62K May 15 18:54 ReachabilityCardinality.xml
-rw-r--r-- 1 mcc users 112 May 15 18:54 ReachabilityDeadlock.txt
-rw-r--r-- 1 mcc users 350 May 15 18:54 ReachabilityDeadlock.xml
-rw-r--r-- 1 mcc users 232K May 15 18:54 ReachabilityFireability.txt
-rw-r--r-- 1 mcc users 873K May 15 18:54 ReachabilityFireability.xml
-rw-r--r-- 1 mcc users 4.7K May 15 18:54 UpperBounds.txt
-rw-r--r-- 1 mcc users 9.9K May 15 18:54 UpperBounds.xml
-rw-r--r-- 1 mcc users 5 May 15 18:50 equiv_col
-rw-r--r-- 1 mcc users 3 May 15 18:50 instance
-rw-r--r-- 1 mcc users 6 May 15 18:50 iscolored
-rw-r--r-- 1 mcc users 2.5M May 15 18:50 model.pnml
=====================================================================
Generated by BenchKit 2-3637
Executing tool itstoolsl
Input is PhilosophersDyn-PT-10, examination is LTLCardinality
Time confinement is 3600 seconds
Memory confinement is 16384 MBytes
Number of cores is 4
Run identifier is r261-csrt-152732586100251
=====================================================================
--------------------
content from stdout:
=== Data for post analysis generated by BenchKit (invocation template)
The expected result is a vector of booleans
BOOL_VECTOR
here is the order used to build the result vector(from text file)
FORMULA_NAME PhilosophersDyn-PT-10-LTLCardinality-00
FORMULA_NAME PhilosophersDyn-PT-10-LTLCardinality-01
FORMULA_NAME PhilosophersDyn-PT-10-LTLCardinality-02
FORMULA_NAME PhilosophersDyn-PT-10-LTLCardinality-03
FORMULA_NAME PhilosophersDyn-PT-10-LTLCardinality-04
FORMULA_NAME PhilosophersDyn-PT-10-LTLCardinality-05
FORMULA_NAME PhilosophersDyn-PT-10-LTLCardinality-06
FORMULA_NAME PhilosophersDyn-PT-10-LTLCardinality-07
FORMULA_NAME PhilosophersDyn-PT-10-LTLCardinality-08
FORMULA_NAME PhilosophersDyn-PT-10-LTLCardinality-09
FORMULA_NAME PhilosophersDyn-PT-10-LTLCardinality-10
FORMULA_NAME PhilosophersDyn-PT-10-LTLCardinality-11
FORMULA_NAME PhilosophersDyn-PT-10-LTLCardinality-12
FORMULA_NAME PhilosophersDyn-PT-10-LTLCardinality-13
FORMULA_NAME PhilosophersDyn-PT-10-LTLCardinality-14
FORMULA_NAME PhilosophersDyn-PT-10-LTLCardinality-15
=== Now, execution of the tool begins
BK_START 1527509951620
Using solver Z3 to compute partial order matrices.
Built C files in :
/home/mcc/execution
Invoking ITS tools like this :CommandLine [args=[/home/mcc/BenchKit/itstools/plugins/fr.lip6.move.gal.itstools.binaries_1.0.0.201805241334/bin/its-ltl-linux64, --gc-threshold, 2000000, -i, /home/mcc/execution/LTLCardinality.pnml.gal, -t, CGAL, -LTL, /home/mcc/execution/LTLCardinality.ltl, -c, -stutter-deadlock], workingDir=/home/mcc/execution]
its-ltl command run as :
/home/mcc/BenchKit/itstools/plugins/fr.lip6.move.gal.itstools.binaries_1.0.0.201805241334/bin/its-ltl-linux64 --gc-threshold 2000000 -i /home/mcc/execution/LTLCardinality.pnml.gal -t CGAL -LTL /home/mcc/execution/LTLCardinality.ltl -c -stutter-deadlock
Read 16 LTL properties
Checking formula 0 : !(((("((((((((((Think_9+Think_8)+Think_1)+Think_2)+Think_3)+Think_7)+Think_6)+Think_4)+Think_5)+Think_10)>=1)")U("((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((Neighbourhood_7_3+Neighbourhood_2_10)+Neighbourhood_9_1)+Neighbourhood_3_8)+Neighbourhood_8_2)+Neighbourhood_5_5)+Neighbourhood_8_9)+Neighbourhood_3_4)+Neighbourhood_7_7)+Neighbourhood_3_6)+Neighbourhood_4_8)+Neighbourhood_5_10)+Neighbourhood_5_1)+Neighbourhood_6_5)+Neighbourhood_1_9)+Neighbourhood_1_7)+Neighbourhood_6_3)+Neighbourhood_9_2)+Neighbourhood_2_2)+Neighbourhood_10_4)+Neighbourhood_7_5)+Neighbourhood_1_2)+Neighbourhood_5_8)+Neighbourhood_6_1)+Neighbourhood_9_9)+Neighbourhood_2_7)+Neighbourhood_6_7)+Neighbourhood_10_2)+Neighbourhood_4_4)+Neighbourhood_2_6)+Neighbourhood_5_3)+Neighbourhood_4_10)+Neighbourhood_3_1)+Neighbourhood_9_4)+Neighbourhood_9_7)+Neighbourhood_8_5)+Neighbourhood_4_3)+Neighbourhood_6_8)+Neighbourhood_9_6)+Neighbourhood_10_8)+Neighbourhood_1_6)+Neighbourhood_9_8)+Neighbourhood_1_4)+Neighbourhood_6_9)+Neighbourhood_7_10)+Neighbourhood_8_7)+Neighbourhood_7_8)+Neighbourhood_4_1)+Neighbourhood_6_10)+Neighbourhood_4_2)+Neighbourhood_3_3)+Neighbourhood_1_5)+Neighbourhood_2_4)+Neighbourhood_8_8)+Neighbourhood_7_9)+Neighbourhood_10_6)+Neighbourhood_10_5)+Neighbourhood_3_2)+Neighbourhood_9_3)+Neighbourhood_4_7)+Neighbourhood_6_4)+Neighbourhood_2_3)+Neighbourhood_5_2)+Neighbourhood_6_6)+Neighbourhood_8_1)+Neighbourhood_10_3)+Neighbourhood_2_5)+Neighbourhood_1_1)+Neighbourhood_10_7)+Neighbourhood_1_3)+Neighbourhood_9_5)+Neighbourhood_7_1)+Neighbourhood_3_7)+Neighbourhood_8_6)+Neighbourhood_6_2)+Neighbourhood_5_4)+Neighbourhood_8_10)+Neighbourhood_5_9)+Neighbourhood_9_10)+Neighbourhood_4_9)+Neighbourhood_7_6)+Neighbourhood_2_1)+Neighbourhood_1_8)+Neighbourhood_3_5)+Neighbourhood_10_9)+Neighbourhood_3_9)+Neighbourhood_8_4)+Neighbourhood_5_6)+Neighbourhood_10_1)+Neighbourhood_7_2)+Neighbourhood_4_5)+Neighbourhood_2_8)+Neighbourhood_10_10)+Neighbourhood_4_6)+Neighbourhood_7_4)+Neighbourhood_8_3)+Neighbourhood_1_10)+Neighbourhood_2_9)+Neighbourhood_5_7)+Neighbourhood_3_10)<=(((((((((HasLeft_6+HasLeft_1)+HasLeft_8)+HasLeft_9)+HasLeft_3)+HasLeft_10)+HasLeft_2)+HasLeft_5)+HasLeft_4)+HasLeft_7))"))U(F(G("((((((((((Forks_10+Forks_8)+Forks_1)+Forks_7)+Forks_6)+Forks_5)+Forks_4)+Forks_3)+Forks_9)+Forks_2)>=2)")))))
Formula 0 simplified : !(("((((((((((Think_9+Think_8)+Think_1)+Think_2)+Think_3)+Think_7)+Think_6)+Think_4)+Think_5)+Think_10)>=1)" U "((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((((Neighbourhood_7_3+Neighbourhood_2_10)+Neighbourhood_9_1)+Neighbourhood_3_8)+Neighbourhood_8_2)+Neighbourhood_5_5)+Neighbourhood_8_9)+Neighbourhood_3_4)+Neighbourhood_7_7)+Neighbourhood_3_6)+Neighbourhood_4_8)+Neighbourhood_5_10)+Neighbourhood_5_1)+Neighbourhood_6_5)+Neighbourhood_1_9)+Neighbourhood_1_7)+Neighbourhood_6_3)+Neighbourhood_9_2)+Neighbourhood_2_2)+Neighbourhood_10_4)+Neighbourhood_7_5)+Neighbourhood_1_2)+Neighbourhood_5_8)+Neighbourhood_6_1)+Neighbourhood_9_9)+Neighbourhood_2_7)+Neighbourhood_6_7)+Neighbourhood_10_2)+Neighbourhood_4_4)+Neighbourhood_2_6)+Neighbourhood_5_3)+Neighbourhood_4_10)+Neighbourhood_3_1)+Neighbourhood_9_4)+Neighbourhood_9_7)+Neighbourhood_8_5)+Neighbourhood_4_3)+Neighbourhood_6_8)+Neighbourhood_9_6)+Neighbourhood_10_8)+Neighbourhood_1_6)+Neighbourhood_9_8)+Neighbourhood_1_4)+Neighbourhood_6_9)+Neighbourhood_7_10)+Neighbourhood_8_7)+Neighbourhood_7_8)+Neighbourhood_4_1)+Neighbourhood_6_10)+Neighbourhood_4_2)+Neighbourhood_3_3)+Neighbourhood_1_5)+Neighbourhood_2_4)+Neighbourhood_8_8)+Neighbourhood_7_9)+Neighbourhood_10_6)+Neighbourhood_10_5)+Neighbourhood_3_2)+Neighbourhood_9_3)+Neighbourhood_4_7)+Neighbourhood_6_4)+Neighbourhood_2_3)+Neighbourhood_5_2)+Neighbourhood_6_6)+Neighbourhood_8_1)+Neighbourhood_10_3)+Neighbourhood_2_5)+Neighbourhood_1_1)+Neighbourhood_10_7)+Neighbourhood_1_3)+Neighbourhood_9_5)+Neighbourhood_7_1)+Neighbourhood_3_7)+Neighbourhood_8_6)+Neighbourhood_6_2)+Neighbourhood_5_4)+Neighbourhood_8_10)+Neighbourhood_5_9)+Neighbourhood_9_10)+Neighbourhood_4_9)+Neighbourhood_7_6)+Neighbourhood_2_1)+Neighbourhood_1_8)+Neighbourhood_3_5)+Neighbourhood_10_9)+Neighbourhood_3_9)+Neighbourhood_8_4)+Neighbourhood_5_6)+Neighbourhood_10_1)+Neighbourhood_7_2)+Neighbourhood_4_5)+Neighbourhood_2_8)+Neighbourhood_10_10)+Neighbourhood_4_6)+Neighbourhood_7_4)+Neighbourhood_8_3)+Neighbourhood_1_10)+Neighbourhood_2_9)+Neighbourhood_5_7)+Neighbourhood_3_10)<=(((((((((HasLeft_6+HasLeft_1)+HasLeft_8)+HasLeft_9)+HasLeft_3)+HasLeft_10)+HasLeft_2)+HasLeft_5)+HasLeft_4)+HasLeft_7))") U FG"((((((((((Forks_10+Forks_8)+Forks_1)+Forks_7)+Forks_6)+Forks_5)+Forks_4)+Forks_3)+Forks_9)+Forks_2)>=2)")
Running compilation step : CommandLine [args=[gcc, -c, -I/home/mcc/BenchKit//lts_install_dir//include, -I., -std=c99, -fPIC, -O3, model.c], workingDir=/home/mcc/execution]
Compilation finished in 30082 ms.
Running link step : CommandLine [args=[gcc, -shared, -o, gal.so, model.o], workingDir=/home/mcc/execution]
Link finished in 146 ms.
Running LTSmin : CommandLine [args=[/home/mcc/BenchKit//lts_install_dir//bin/pins2lts-mc, ./gal.so, --threads=1, -p, --pins-guards, --when, --ltl, (((LTLAP0==true))U((LTLAP1==true)))U(<>([]((LTLAP2==true)))), --buchi-type=spotba], workingDir=/home/mcc/execution]
LTSmin run took 14145 ms.
FORMULA PhilosophersDyn-PT-10-LTLCardinality-00 FALSE TECHNIQUES PARTIAL_ORDER EXPLICIT LTSMIN SAT_SMT
Running LTSmin : CommandLine [args=[/home/mcc/BenchKit//lts_install_dir//bin/pins2lts-mc, ./gal.so, --threads=1, -p, --pins-guards, --when, --ltl, (LTLAP3==true), --buchi-type=spotba], workingDir=/home/mcc/execution]
LTSmin run took 14072 ms.
FORMULA PhilosophersDyn-PT-10-LTLCardinality-01 FALSE TECHNIQUES PARTIAL_ORDER EXPLICIT LTSMIN SAT_SMT
Running LTSmin : CommandLine [args=[/home/mcc/BenchKit//lts_install_dir//bin/pins2lts-mc, ./gal.so, --threads=1, --when, --ltl, X([](<>([]((LTLAP4==true))))), --buchi-type=spotba], workingDir=/home/mcc/execution]
LTSmin run took 272 ms.
FORMULA PhilosophersDyn-PT-10-LTLCardinality-02 FALSE TECHNIQUES PARTIAL_ORDER EXPLICIT LTSMIN SAT_SMT
Running LTSmin : CommandLine [args=[/home/mcc/BenchKit//lts_install_dir//bin/pins2lts-mc, ./gal.so, --threads=1, -p, --pins-guards, --when, --ltl, (((LTLAP5==true))U((LTLAP6==true)))U((LTLAP7==true)), --buchi-type=spotba], workingDir=/home/mcc/execution]
LTSmin run took 14084 ms.
FORMULA PhilosophersDyn-PT-10-LTLCardinality-03 FALSE TECHNIQUES PARTIAL_ORDER EXPLICIT LTSMIN SAT_SMT
Running LTSmin : CommandLine [args=[/home/mcc/BenchKit//lts_install_dir//bin/pins2lts-mc, ./gal.so, --threads=1, -p, --pins-guards, --when, --ltl, <>((LTLAP8==true)), --buchi-type=spotba], workingDir=/home/mcc/execution]
LTSmin run took 14001 ms.
FORMULA PhilosophersDyn-PT-10-LTLCardinality-04 TRUE TECHNIQUES PARTIAL_ORDER EXPLICIT LTSMIN SAT_SMT
Running LTSmin : CommandLine [args=[/home/mcc/BenchKit//lts_install_dir//bin/pins2lts-mc, ./gal.so, --threads=1, -p, --pins-guards, --when, --ltl, ([]((LTLAP9==true)))U((LTLAP10==true)), --buchi-type=spotba], workingDir=/home/mcc/execution]
LTSmin run took 14117 ms.
FORMULA PhilosophersDyn-PT-10-LTLCardinality-05 TRUE TECHNIQUES PARTIAL_ORDER EXPLICIT LTSMIN SAT_SMT
Running LTSmin : CommandLine [args=[/home/mcc/BenchKit//lts_install_dir//bin/pins2lts-mc, ./gal.so, --threads=1, --when, --ltl, X((LTLAP11==true)), --buchi-type=spotba], workingDir=/home/mcc/execution]
LTSmin run took 89 ms.
FORMULA PhilosophersDyn-PT-10-LTLCardinality-06 TRUE TECHNIQUES PARTIAL_ORDER EXPLICIT LTSMIN SAT_SMT
Running LTSmin : CommandLine [args=[/home/mcc/BenchKit//lts_install_dir//bin/pins2lts-mc, ./gal.so, --threads=1, -p, --pins-guards, --when, --ltl, <>(<>([]([]((LTLAP4==true))))), --buchi-type=spotba], workingDir=/home/mcc/execution]
LTSmin run took 15568 ms.
FORMULA PhilosophersDyn-PT-10-LTLCardinality-07 FALSE TECHNIQUES PARTIAL_ORDER EXPLICIT LTSMIN SAT_SMT
Running LTSmin : CommandLine [args=[/home/mcc/BenchKit//lts_install_dir//bin/pins2lts-mc, ./gal.so, --threads=1, --when, --ltl, (X(X((LTLAP12==true))))U(<>((LTLAP13==true))), --buchi-type=spotba], workingDir=/home/mcc/execution]
LTSmin run took 60 ms.
FORMULA PhilosophersDyn-PT-10-LTLCardinality-08 TRUE TECHNIQUES PARTIAL_ORDER EXPLICIT LTSMIN SAT_SMT
Running LTSmin : CommandLine [args=[/home/mcc/BenchKit//lts_install_dir//bin/pins2lts-mc, ./gal.so, --threads=1, -p, --pins-guards, --when, --ltl, ([](<>((LTLAP14==true))))U(((LTLAP15==true))U((LTLAP16==true))), --buchi-type=spotba], workingDir=/home/mcc/execution]
LTSmin run took 14207 ms.
FORMULA PhilosophersDyn-PT-10-LTLCardinality-09 TRUE TECHNIQUES PARTIAL_ORDER EXPLICIT LTSMIN SAT_SMT
Running LTSmin : CommandLine [args=[/home/mcc/BenchKit//lts_install_dir//bin/pins2lts-mc, ./gal.so, --threads=1, --when, --ltl, <>(X(<>(<>((LTLAP17==true))))), --buchi-type=spotba], workingDir=/home/mcc/execution]
LTSmin run took 16097 ms.
FORMULA PhilosophersDyn-PT-10-LTLCardinality-10 TRUE TECHNIQUES PARTIAL_ORDER EXPLICIT LTSMIN SAT_SMT
Running LTSmin : CommandLine [args=[/home/mcc/BenchKit//lts_install_dir//bin/pins2lts-mc, ./gal.so, --threads=1, --when, --ltl, X(X((LTLAP18==true))), --buchi-type=spotba], workingDir=/home/mcc/execution]
LTSmin run took 137 ms.
FORMULA PhilosophersDyn-PT-10-LTLCardinality-11 FALSE TECHNIQUES PARTIAL_ORDER EXPLICIT LTSMIN SAT_SMT
Running LTSmin : CommandLine [args=[/home/mcc/BenchKit//lts_install_dir//bin/pins2lts-mc, ./gal.so, --threads=1, --when, --ltl, [](<>(X([]((LTLAP19==true))))), --buchi-type=spotba], workingDir=/home/mcc/execution]
LTSmin run took 395 ms.
FORMULA PhilosophersDyn-PT-10-LTLCardinality-12 FALSE TECHNIQUES PARTIAL_ORDER EXPLICIT LTSMIN SAT_SMT
Running LTSmin : CommandLine [args=[/home/mcc/BenchKit//lts_install_dir//bin/pins2lts-mc, ./gal.so, --threads=1, -p, --pins-guards, --when, --ltl, (LTLAP20==true), --buchi-type=spotba], workingDir=/home/mcc/execution]
LTSmin run took 13983 ms.
FORMULA PhilosophersDyn-PT-10-LTLCardinality-13 TRUE TECHNIQUES PARTIAL_ORDER EXPLICIT LTSMIN SAT_SMT
Running LTSmin : CommandLine [args=[/home/mcc/BenchKit//lts_install_dir//bin/pins2lts-mc, ./gal.so, --threads=1, -p, --pins-guards, --when, --ltl, <>((LTLAP21==true)), --buchi-type=spotba], workingDir=/home/mcc/execution]
LTSmin run took 14071 ms.
FORMULA PhilosophersDyn-PT-10-LTLCardinality-14 TRUE TECHNIQUES PARTIAL_ORDER EXPLICIT LTSMIN SAT_SMT
Running LTSmin : CommandLine [args=[/home/mcc/BenchKit//lts_install_dir//bin/pins2lts-mc, ./gal.so, --threads=1, --when, --ltl, [](X(X((LTLAP22==true)))), --buchi-type=spotba], workingDir=/home/mcc/execution]
LTSmin run took 734 ms.
FORMULA PhilosophersDyn-PT-10-LTLCardinality-15 FALSE TECHNIQUES PARTIAL_ORDER EXPLICIT LTSMIN SAT_SMT
ITS tools runner thread asked to quit. Dying gracefully.
BK_STOP 1527510134460
--------------------
content from stderr:
+ export BINDIR=/home/mcc/BenchKit/
+ BINDIR=/home/mcc/BenchKit/
++ pwd
+ export MODEL=/home/mcc/execution
+ MODEL=/home/mcc/execution
+ /home/mcc/BenchKit//runeclipse.sh /home/mcc/execution LTLCardinality -its -ltsminpath /home/mcc/BenchKit//lts_install_dir/ -louvain -smt
+ ulimit -s 65536
+ [[ -z '' ]]
+ export LTSMIN_MEM_SIZE=8589934592
+ LTSMIN_MEM_SIZE=8589934592
+ /home/mcc/BenchKit//itstools/its-tools -consoleLog -data /home/mcc/execution/workspace -pnfolder /home/mcc/execution -examination LTLCardinality -z3path /home/mcc/BenchKit//z3/bin/z3 -yices2path /home/mcc/BenchKit//yices/bin/yices -its -ltsminpath /home/mcc/BenchKit//lts_install_dir/ -louvain -smt -vmargs -Dosgi.locking=none -Declipse.stateSaveDelayInterval=-1 -Dosgi.configuration.area=/tmp/.eclipse -Xss8m -Xms40m -Xmx8192m -Dfile.encoding=UTF-8 -Dosgi.requiredJavaVersion=1.6
May 28, 2018 12:19:14 PM fr.lip6.move.gal.application.Application start
INFO: Running its-tools with arguments : [-pnfolder, /home/mcc/execution, -examination, LTLCardinality, -z3path, /home/mcc/BenchKit//z3/bin/z3, -yices2path, /home/mcc/BenchKit//yices/bin/yices, -its, -ltsminpath, /home/mcc/BenchKit//lts_install_dir/, -louvain, -smt]
May 28, 2018 12:19:14 PM fr.lip6.move.gal.application.MccTranslator transformPNML
INFO: Parsing pnml file : /home/mcc/execution/model.pnml
May 28, 2018 12:19:14 PM fr.lip6.move.gal.nupn.PTNetReader loadFromXML
INFO: Load time of PNML (sax parser for PT used): 300 ms
May 28, 2018 12:19:14 PM fr.lip6.move.gal.pnml.togal.PTGALTransformer handlePage
INFO: Transformed 170 places.
May 28, 2018 12:19:14 PM fr.lip6.move.gal.pnml.togal.PTGALTransformer handlePage
INFO: Transformed 2310 transitions.
May 28, 2018 12:19:15 PM fr.lip6.move.gal.instantiate.GALRewriter flatten
INFO: Flatten gal took : 861 ms
May 28, 2018 12:19:16 PM fr.lip6.move.serialization.SerializationUtil systemToFile
INFO: Time to serialize gal into /home/mcc/execution/LTLCardinality.pnml.gal : 111 ms
May 28, 2018 12:19:16 PM fr.lip6.move.serialization.SerializationUtil serializePropertiesForITSLTLTools
INFO: Time to serialize properties into /home/mcc/execution/LTLCardinality.ltl : 1 ms
May 28, 2018 12:19:16 PM fr.lip6.move.gal.semantics.DeterministicNextBuilder getDeterministicNext
INFO: Input system was already deterministic with 2310 transitions.
May 28, 2018 12:19:16 PM fr.lip6.move.gal.gal2pins.Gal2PinsTransformerNext transform
INFO: Too many transitions (2310) to apply POR reductions. Disabling POR matrices.
May 28, 2018 12:19:17 PM fr.lip6.move.gal.gal2pins.Gal2PinsTransformerNext transform
INFO: Built C files in 1176ms conformant to PINS in folder :/home/mcc/execution
Sequence of Actions to be Executed by the VM
This is useful if one wants to reexecute the tool in the VM from the submitted image disk.
set -x
# this is for BenchKit: configuration of major elements for the test
export BK_INPUT="PhilosophersDyn-PT-10"
export BK_EXAMINATION="LTLCardinality"
export BK_TOOL="itstoolsl"
export BK_RESULT_DIR="/tmp/BK_RESULTS/OUTPUTS"
export BK_TIME_CONFINEMENT="3600"
export BK_MEMORY_CONFINEMENT="16384"
# this is specific to your benchmark or test
export BIN_DIR="$HOME/BenchKit/bin"
# remove the execution directoty if it exists (to avoid increse of .vmdk images)
if [ -d execution ] ; then
rm -rf execution
fi
tar xzf /home/mcc/BenchKit/INPUTS/PhilosophersDyn-PT-10.tgz
mv PhilosophersDyn-PT-10 execution
cd execution
pwd
ls -lh
# this is for BenchKit: explicit launching of the test
echo "====================================================================="
echo " Generated by BenchKit 2-3637"
echo " Executing tool itstoolsl"
echo " Input is PhilosophersDyn-PT-10, examination is LTLCardinality"
echo " Time confinement is $BK_TIME_CONFINEMENT seconds"
echo " Memory confinement is 16384 MBytes"
echo " Number of cores is 4"
echo " Run identifier is r261-csrt-152732586100251"
echo "====================================================================="
echo
echo "--------------------"
echo "content from stdout:"
echo
echo "=== Data for post analysis generated by BenchKit (invocation template)"
echo
if [ "LTLCardinality" = "UpperBounds" ] ; then
echo "The expected result is a vector of positive values"
echo NUM_VECTOR
elif [ "LTLCardinality" != "StateSpace" ] ; then
echo "The expected result is a vector of booleans"
echo BOOL_VECTOR
else
echo "no data necessary for post analysis"
fi
echo
if [ -f "LTLCardinality.txt" ] ; then
echo "here is the order used to build the result vector(from text file)"
for x in $(grep Property LTLCardinality.txt | cut -d ' ' -f 2 | sort -u) ; do
echo "FORMULA_NAME $x"
done
elif [ -f "LTLCardinality.xml" ] ; then # for cunf (txt files deleted;-)
echo echo "here is the order used to build the result vector(from xml file)"
for x in $(grep '
echo "FORMULA_NAME $x"
done
fi
echo
echo "=== Now, execution of the tool begins"
echo
echo -n "BK_START "
date -u +%s%3N
echo
timeout -s 9 $BK_TIME_CONFINEMENT bash -c "/home/mcc/BenchKit/BenchKit_head.sh 2> STDERR ; echo ; echo -n \"BK_STOP \" ; date -u +%s%3N"
if [ $? -eq 137 ] ; then
echo
echo "BK_TIME_CONFINEMENT_REACHED"
fi
echo
echo "--------------------"
echo "content from stderr:"
echo
cat STDERR ;