context: SyFgets | |
observation: 0 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684358144.0 | |
reward: 0.0 | |
observation: 1 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684358656.0 | |
reward: 0.0 | |
observation: 2 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684357888.0 | |
reward: 0.0 | |
observation: 3 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684358144.0 | |
reward: 0.0 | |
observation: 4 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 3758100736.0 | |
reward: 0.0 | |
observation: 5 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 3758096896.0 | |
reward: 0.0 | |
observation: 6 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2147485184.0 | |
reward: 0.0 | |
observation: 7 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2147484928.0 | |
reward: 0.0 | |
observation: 8 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684358144.0 | |
reward: 0.0 | |
observation: 9 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2147484928.0 | |
reward: 0.0 | |
observation: 10 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2147484928.0 | |
reward: 0.0 | |
observation: 11 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684354560.0 | |
reward: 0.0 | |
observation: 12 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684354560.0 | |
reward: 0.0 | |
observation: 13 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684357632.0 | |
reward: 0.0 | |
observation: 14 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684354560.0 | |
reward: 0.0 | |
observation: 15 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684354560.0 | |
reward: 0.0 | |
observation: 16 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2147484672.0 | |
reward: 0.0 | |
observation: 17 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684357632.0 | |
reward: 0.0 | |
observation: 18 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684354560.0 | |
reward: 0.0 | |
observation: 19 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684357632.0 | |
reward: 0.0 | |
observation: 20 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684357632.0 | |
reward: 0.0 | |
observation: 21 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2147483904.0 | |
reward: 0.0 | |
observation: 22 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2147483904.0 | |
reward: 0.0 | |
observation: 23 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2147483904.0 | |
reward: 0.0 | |
observation: 24 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2147483904.0 | |
reward: 0.0 | |
observation: 25 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 3221225472.0 | |
reward: 0.0 | |
observation: 26 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684354816.0 | |
reward: 0.0 | |
observation: 27 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2147483648.0 | |
reward: 0.0 | |
observation: 28 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2147484928.0 | |
reward: 0.0 | |
observation: 29 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2147484928.0 | |
reward: 0.0 | |
observation: 30 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684354560.0 | |
reward: 0.0 | |
observation: 31 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684355840.0 | |
reward: 0.0 | |
observation: 32 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684354816.0 | |
reward: 0.0 | |
observation: 33 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684354816.0 | |
reward: 0.0 | |
observation: 34 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684357888.0 | |
reward: 0.0 | |
observation: 35 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684355840.0 | |
reward: 0.0 | |
observation: 36 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684354560.0 | |
reward: 0.0 | |
observation: 37 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 3758096896.0 | |
reward: 0.0 | |
observation: 38 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684358144.0 | |
reward: 0.0 | |
observation: 39 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 3534.0 | |
reward: 0.0 | |
observation: 40 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684358144.0 | |
reward: 0.0 | |
observation: 41 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 3518.0 | |
reward: 0.0 | |
observation: 42 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 3022.0 | |
reward: 0.0 | |
observation: 43 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 3006.0 | |
reward: 0.0 | |
observation: 44 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684358656.0 | |
reward: 0.0 | |
observation: 45 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 4046.0 | |
reward: 0.0 | |
observation: 46 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 3758099712.0 | |
reward: 0.0 | |
observation: 47 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 3758100736.0 | |
reward: 0.0 | |
observation: 48 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 4304.0 | |
reward: 0.0 | |
observation: 49 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684358144.0 | |
reward: 0.0 | |
observation: 50 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 3610.0 | |
reward: 0.0 | |
observation: 51 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684357888.0 | |
reward: 0.0 | |
observation: 52 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 3758097152.0 | |
reward: 0.0 | |
observation: 53 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684354816.0 | |
reward: 0.0 | |
observation: 54 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684354560.0 | |
reward: 0.0 | |
observation: 55 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 3758097152.0 | |
reward: 0.0 | |
observation: 56 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684358144.0 | |
reward: 0.0 | |
observation: 57 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684354816.0 | |
reward: 0.0 | |
observation: 58 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684354816.0 | |
reward: 0.0 | |
observation: 59 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684357888.0 | |
reward: 0.0 | |
observation: 60 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684354816.0 | |
reward: 0.0 | |
observation: 61 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684358144.0 | |
reward: 0.0 | |
observation: 62 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2147484928.0 | |
reward: 0.0 | |
observation: 63 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684354560.0 | |
reward: 0.0 | |
observation: 64 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684358144.0 | |
reward: 0.0 | |
observation: 65 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2147484928.0 | |
reward: 0.0 | |
observation: 66 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684354560.0 | |
reward: 0.0 | |
observation: 67 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684354560.0 | |
reward: 0.0 | |
observation: 68 | |
li_size: 0 | |
stage: 0 | |
weight: 0.0 | |
priority: 2684354560.0 | |
reward: 37.06101608276367 |