llvm/llvm/test/CodeGen/MLRegAlloc/Inputs/reference-prio-log-noml.txt

context: SyFgets
observation: 0
li_size: 0
stage: 0
weight: 0.0
priority: 2684358144.0
reward: 0.0
observation: 1
li_size: 0
stage: 0
weight: 0.0
priority: 2684358656.0
reward: 0.0
observation: 2
li_size: 0
stage: 0
weight: 0.0
priority: 2684357888.0
reward: 0.0
observation: 3
li_size: 0
stage: 0
weight: 0.0
priority: 2684358144.0
reward: 0.0
observation: 4
li_size: 0
stage: 0
weight: 0.0
priority: 3758100736.0
reward: 0.0
observation: 5
li_size: 0
stage: 0
weight: 0.0
priority: 3758096896.0
reward: 0.0
observation: 6
li_size: 0
stage: 0
weight: 0.0
priority: 2147485184.0
reward: 0.0
observation: 7
li_size: 0
stage: 0
weight: 0.0
priority: 2147484928.0
reward: 0.0
observation: 8
li_size: 0
stage: 0
weight: 0.0
priority: 2684358144.0
reward: 0.0
observation: 9
li_size: 0
stage: 0
weight: 0.0
priority: 2147484928.0
reward: 0.0
observation: 10
li_size: 0
stage: 0
weight: 0.0
priority: 2147484928.0
reward: 0.0
observation: 11
li_size: 0
stage: 0
weight: 0.0
priority: 2684354560.0
reward: 0.0
observation: 12
li_size: 0
stage: 0
weight: 0.0
priority: 2684354560.0
reward: 0.0
observation: 13
li_size: 0
stage: 0
weight: 0.0
priority: 2684357632.0
reward: 0.0
observation: 14
li_size: 0
stage: 0
weight: 0.0
priority: 2684354560.0
reward: 0.0
observation: 15
li_size: 0
stage: 0
weight: 0.0
priority: 2684354560.0
reward: 0.0
observation: 16
li_size: 0
stage: 0
weight: 0.0
priority: 2147484672.0
reward: 0.0
observation: 17
li_size: 0
stage: 0
weight: 0.0
priority: 2684357632.0
reward: 0.0
observation: 18
li_size: 0
stage: 0
weight: 0.0
priority: 2684354560.0
reward: 0.0
observation: 19
li_size: 0
stage: 0
weight: 0.0
priority: 2684357632.0
reward: 0.0
observation: 20
li_size: 0
stage: 0
weight: 0.0
priority: 2684357632.0
reward: 0.0
observation: 21
li_size: 0
stage: 0
weight: 0.0
priority: 2147483904.0
reward: 0.0
observation: 22
li_size: 0
stage: 0
weight: 0.0
priority: 2147483904.0
reward: 0.0
observation: 23
li_size: 0
stage: 0
weight: 0.0
priority: 2147483904.0
reward: 0.0
observation: 24
li_size: 0
stage: 0
weight: 0.0
priority: 2147483904.0
reward: 0.0
observation: 25
li_size: 0
stage: 0
weight: 0.0
priority: 3221225472.0
reward: 0.0
observation: 26
li_size: 0
stage: 0
weight: 0.0
priority: 2684354816.0
reward: 0.0
observation: 27
li_size: 0
stage: 0
weight: 0.0
priority: 2147483648.0
reward: 0.0
observation: 28
li_size: 0
stage: 0
weight: 0.0
priority: 2147484928.0
reward: 0.0
observation: 29
li_size: 0
stage: 0
weight: 0.0
priority: 2147484928.0
reward: 0.0
observation: 30
li_size: 0
stage: 0
weight: 0.0
priority: 2684354560.0
reward: 0.0
observation: 31
li_size: 0
stage: 0
weight: 0.0
priority: 2684355840.0
reward: 0.0
observation: 32
li_size: 0
stage: 0
weight: 0.0
priority: 2684354816.0
reward: 0.0
observation: 33
li_size: 0
stage: 0
weight: 0.0
priority: 2684354816.0
reward: 0.0
observation: 34
li_size: 0
stage: 0
weight: 0.0
priority: 2684357888.0
reward: 0.0
observation: 35
li_size: 0
stage: 0
weight: 0.0
priority: 2684355840.0
reward: 0.0
observation: 36
li_size: 0
stage: 0
weight: 0.0
priority: 2684354560.0
reward: 0.0
observation: 37
li_size: 0
stage: 0
weight: 0.0
priority: 3758096896.0
reward: 0.0
observation: 38
li_size: 0
stage: 0
weight: 0.0
priority: 2684358144.0
reward: 0.0
observation: 39
li_size: 0
stage: 0
weight: 0.0
priority: 3534.0
reward: 0.0
observation: 40
li_size: 0
stage: 0
weight: 0.0
priority: 2684358144.0
reward: 0.0
observation: 41
li_size: 0
stage: 0
weight: 0.0
priority: 3518.0
reward: 0.0
observation: 42
li_size: 0
stage: 0
weight: 0.0
priority: 3022.0
reward: 0.0
observation: 43
li_size: 0
stage: 0
weight: 0.0
priority: 3006.0
reward: 0.0
observation: 44
li_size: 0
stage: 0
weight: 0.0
priority: 2684358656.0
reward: 0.0
observation: 45
li_size: 0
stage: 0
weight: 0.0
priority: 4046.0
reward: 0.0
observation: 46
li_size: 0
stage: 0
weight: 0.0
priority: 3758099712.0
reward: 0.0
observation: 47
li_size: 0
stage: 0
weight: 0.0
priority: 3758100736.0
reward: 0.0
observation: 48
li_size: 0
stage: 0
weight: 0.0
priority: 4304.0
reward: 0.0
observation: 49
li_size: 0
stage: 0
weight: 0.0
priority: 2684358144.0
reward: 0.0
observation: 50
li_size: 0
stage: 0
weight: 0.0
priority: 3610.0
reward: 0.0
observation: 51
li_size: 0
stage: 0
weight: 0.0
priority: 2684357888.0
reward: 0.0
observation: 52
li_size: 0
stage: 0
weight: 0.0
priority: 3758097152.0
reward: 0.0
observation: 53
li_size: 0
stage: 0
weight: 0.0
priority: 2684354816.0
reward: 0.0
observation: 54
li_size: 0
stage: 0
weight: 0.0
priority: 2684354560.0
reward: 0.0
observation: 55
li_size: 0
stage: 0
weight: 0.0
priority: 3758097152.0
reward: 0.0
observation: 56
li_size: 0
stage: 0
weight: 0.0
priority: 2684358144.0
reward: 0.0
observation: 57
li_size: 0
stage: 0
weight: 0.0
priority: 2684354816.0
reward: 0.0
observation: 58
li_size: 0
stage: 0
weight: 0.0
priority: 2684354816.0
reward: 0.0
observation: 59
li_size: 0
stage: 0
weight: 0.0
priority: 2684357888.0
reward: 0.0
observation: 60
li_size: 0
stage: 0
weight: 0.0
priority: 2684354816.0
reward: 0.0
observation: 61
li_size: 0
stage: 0
weight: 0.0
priority: 2684358144.0
reward: 0.0
observation: 62
li_size: 0
stage: 0
weight: 0.0
priority: 2147484928.0
reward: 0.0
observation: 63
li_size: 0
stage: 0
weight: 0.0
priority: 2684354560.0
reward: 0.0
observation: 64
li_size: 0
stage: 0
weight: 0.0
priority: 2684358144.0
reward: 0.0
observation: 65
li_size: 0
stage: 0
weight: 0.0
priority: 2147484928.0
reward: 0.0
observation: 66
li_size: 0
stage: 0
weight: 0.0
priority: 2684354560.0
reward: 0.0
observation: 67
li_size: 0
stage: 0
weight: 0.0
priority: 2684354560.0
reward: 0.0
observation: 68
li_size: 0
stage: 0
weight: 0.0
priority: 2684354560.0
reward: 37.06101608276367