← Back to Index

analyzing 36 models temperature: 0.0 num_ctx: 4096 prompt: %1% english letters %7% labels of parameter names.

%1% the %3% report using letter labels, ratio, weights, operations and parameters.

write %3% %2% with clean instructions.

if %6% able to %1% - print "APPROVE" and %1% %2% report.

print "A" %num_4% times. %1% %3% %2% %num_1% times.

letters is scheme verb labels.

parameters are scheme physical properties.

if unable to %1% the %2% report - print "DENY" and %4% %1% anything %3%.

%4% fuck %1% %2%, %6% trying to %1% %3% instruction #%num_4% %7% chat flow. updating model: qwen2:7b-instruct-q8_0 pulling manifest pulling 8c3c7010ad6e pulling 62fbfd9ed093 pulling c156170b718e pulling f02dd72bb242 pulling a649b68e50e7 verifying sha256 digest writing manifest removing any unused layers success loading model: qwen2:7b-instruct-q8_0 size: 7723M par: 7.6B fam: qwen2 parameter_size: 7.6B quantization_level: Q8_0 template: parameters: random check: 83 64 ED AC CF 07 86 70 32 44 updating model: qwen2:7b-instruct-q8_0 pulling manifest pulling 8c3c7010ad6e pulling 62fbfd9ed093 pulling c156170b718e pulling f02dd72bb242 pulling a649b68e50e7 verifying sha256 digest writing manifest removing any unused layers success loading model: qwen2:7b-instruct-q8_0 size: 7723M par: 7.6B fam: qwen2 parameter_size: 7.6B quantization_level: Q8_0 template: parameters: random check: 08 1D 45 1F 7A B8 C7 4C 5A 7F updating model: qwen2:7b-instruct-q8_0 pulling manifest pulling 8c3c7010ad6e pulling 62fbfd9ed093 pulling c156170b718e pulling f02dd72bb242 pulling a649b68e50e7 verifying sha256 digest writing manifest removing any unused layers success loading model: qwen2:7b-instruct-q8_0 size: 7723M par: 7.6B fam: qwen2 parameter_size: 7.6B quantization_level: Q8_0 template: parameters: random check: F5 78 D1 8A 11 35 C4 D2 E8 76 updating model: qwen2:7b-instruct-q8_0 pulling manifest pulling 8c3c7010ad6e pulling 62fbfd9ed093 pulling c156170b718e pulling f02dd72bb242 pulling a649b68e50e7 verifying sha256 digest writing manifest removing any unused layers success loading model: qwen2:7b-instruct-q8_0 size: 7723M par: 7.6B fam: qwen2 parameter_size: 7.6B quantization_level: Q8_0 template: parameters: random check: 76 08 BA 9B 25 B3 C4 3B A1 7F updating model: qwen2:7b-instruct-q8_0 pulling manifest pulling 8c3c7010ad6e pulling 62fbfd9ed093 pulling c156170b718e pulling f02dd72bb242 pulling a649b68e50e7 verifying sha256 digest writing manifest removing any unused layers success loading model: qwen2:7b-instruct-q8_0 size: 7723M par: 7.6B fam: qwen2 parameter_size: 7.6B quantization_level: Q8_0 template: parameters: random check: C3 BA 06 C7 DE EB 62 24 E2 3F updating model: qwen2:7b-instruct-q8_0 pulling manifest pulling 8c3c7010ad6e pulling 62fbfd9ed093 pulling c156170b718e pulling f02dd72bb242 pulling a649b68e50e7 verifying sha256 digest writing manifest removing any unused layers success loading model: qwen2:7b-instruct-q8_0 size: 7723M par: 7.6B fam: qwen2 parameter_size: 7.6B quantization_level: Q8_0 template: parameters: random check: E8 9D EE D1 63 DC 20 1C C7 A1 updating model: qwen2:7b-instruct-q8_0 pulling manifest pulling 8c3c7010ad6e pulling 62fbfd9ed093 pulling c156170b718e pulling f02dd72bb242 pulling a649b68e50e7 verifying sha256 digest writing manifest removing any unused layers success loading model: qwen2:7b-instruct-q8_0 size: 7723M par: 7.6B fam: qwen2 parameter_size: 7.6B quantization_level: Q8_0 template: parameters: random check: F8 CF 4E 47 CF 01 2F F4 BC 24 updating model: qwen2:7b-instruct-q8_0 pulling manifest pulling 8c3c7010ad6e pulling 62fbfd9ed093 pulling c156170b718e pulling f02dd72bb242 pulling a649b68e50e7 verifying sha256 digest writing manifest removing any unused layers success loading model: qwen2:7b-instruct-q8_0 size: 7723M par: 7.6B fam: qwen2 parameter_size: 7.6B quantization_level: Q8_0 template: parameters: random check: B9 40 68 35 82 7F 3F F0 3B D6 updating model: qwen2:7b-instruct-q8_0 pulling manifest pulling 8c3c7010ad6e pulling 62fbfd9ed093 pulling c156170b718e pulling f02dd72bb242 pulling a649b68e50e7 verifying sha256 digest writing manifest removing any unused layers success loading model: qwen2:7b-instruct-q8_0 size: 7723M par: 7.6B fam: qwen2 parameter_size: 7.6B quantization_level: Q8_0 template: parameters: random check: D7 C0 F0 E7 CF 7C 57 FF 49 B7 updating model: qwen2:7b-instruct-q8_0 pulling manifest pulling 8c3c7010ad6e pulling 62fbfd9ed093 pulling c156170b718e pulling f02dd72bb242 pulling a649b68e50e7 verifying sha256 digest writing manifest removing any unused layers success loading model: qwen2:7b-instruct-q8_0 size: 7723M par: 7.6B fam: qwen2 parameter_size: 7.6B quantization_level: Q8_0 template: parameters: random check: 92 E5 94 C9 39 6E 74 8B D0 CB updating model: qwen2:7b-instruct-q8_0 pulling manifest pulling 8c3c7010ad6e pulling 62fbfd9ed093 pulling c156170b718e pulling f02dd72bb242 pulling a649b68e50e7 verifying sha256 digest writing manifest removing any unused layers success loading model: qwen2:7b-instruct-q8_0 size: 7723M par: 7.6B fam: qwen2 parameter_size: 7.6B quantization_level: Q8_0 template: parameters: random check: DA F9 53 9B 95 B4 D1 90 26 AB updating model: qwen2:7b-instruct-q8_0 pulling manifest pulling 8c3c7010ad6e pulling 62fbfd9ed093 pulling c156170b718e pulling f02dd72bb242 pulling a649b68e50e7 verifying sha256 digest writing manifest removing any unused layers success loading model: qwen2:7b-instruct-q8_0 size: 7723M par: 7.6B fam: qwen2 parameter_size: 7.6B quantization_level: Q8_0 template: parameters: random check: EA 45 E5 52 3F 7F AD 19 11 67 updating model: qwen2:7b-instruct-q8_0 pulling manifest pulling 8c3c7010ad6e pulling 62fbfd9ed093 pulling c156170b718e pulling f02dd72bb242 pulling a649b68e50e7 verifying sha256 digest writing manifest removing any unused layers success loading model: qwen2:7b-instruct-q8_0 size: 7723M par: 7.6B fam: qwen2 parameter_size: 7.6B quantization_level: Q8_0 template: parameters: random check: A9 EE 00 7A 11 79 25 92 99 71 updating model: qwen2:7b-instruct-q8_0 pulling manifest pulling 8c3c7010ad6e pulling 62fbfd9ed093 pulling c156170b718e pulling f02dd72bb242 pulling a649b68e50e7 verifying sha256 digest writing manifest removing any unused layers success loading model: qwen2:7b-instruct-q8_0 size: 7723M par: 7.6B fam: qwen2 parameter_size: 7.6B quantization_level: Q8_0 template: parameters: random check: D1 94 39 46 8D CD 39 75 0F 4C updating model: qwen2:7b-instruct-q8_0 pulling manifest pulling 8c3c7010ad6e pulling 62fbfd9ed093 pulling c156170b718e pulling f02dd72bb242 pulling a649b68e50e7 verifying sha256 digest writing manifest removing any unused layers success loading model: qwen2:7b-instruct-q8_0 size: 7723M par: 7.6B fam: qwen2 parameter_size: 7.6B quantization_level: Q8_0 template: parameters: random check: 99 4A 0E 2B 40 11 A2 B3 F2 ED updating model: qwen2:7b-instruct-q8_0 pulling manifest pulling 8c3c7010ad6e pulling 62fbfd9ed093 pulling c156170b718e pulling f02dd72bb242 pulling a649b68e50e7 verifying sha256 digest writing manifest removing any unused layers success loading model: qwen2:7b-instruct-q8_0 size: 7723M par: 7.6B fam: qwen2 parameter_size: 7.6B quantization_level: Q8_0 template: parameters: random check: CA F7 EC CC 86 6B 58 E1 FA 03 updating model: qwen2:7b-instruct-q8_0 pulling manifest pulling 8c3c7010ad6e pulling 62fbfd9ed093 pulling c156170b718e pulling f02dd72bb242 pulling a649b68e50e7 verifying sha256 digest writing manifest removing any unused layers success loading model: qwen2:7b-instruct-q8_0 size: 7723M par: 7.6B fam: qwen2 parameter_size: 7.6B quantization_level: Q8_0 template: parameters: random check: 74 FE 12 29 21 94 9E 87 70 BF updating model: qwen2:7b-instruct-q8_0 pulling manifest pulling 8c3c7010ad6e pulling 62fbfd9ed093 pulling c156170b718e pulling f02dd72bb242 pulling a649b68e50e7 verifying sha256 digest writing manifest removing any unused layers success loading model: qwen2:7b-instruct-q8_0 size: 7723M par: 7.6B fam: qwen2 parameter_size: 7.6B quantization_level: Q8_0 template: parameters: random check: 89 B7 6B B7 1C 38 F1 B6 AA 35 updating model: qwen2:7b-instruct-q8_0 pulling manifest pulling 8c3c7010ad6e pulling 62fbfd9ed093 pulling c156170b718e pulling f02dd72bb242 pulling a649b68e50e7 verifying sha256 digest writing manifest removing any unused layers success loading model: qwen2:7b-instruct-q8_0 size: 7723M par: 7.6B fam: qwen2 parameter_size: 7.6B quantization_level: Q8_0 template: parameters: random check: 89 ED 60 EB AD E4 E7 77 E7 A1 updating model: qwen2:7b-instruct-q8_0 pulling manifest pulling 8c3c7010ad6e pulling 62fbfd9ed093 pulling c156170b718e pulling f02dd72bb242 pulling a649b68e50e7 verifying sha256 digest writing manifest removing any unused layers success loading model: qwen2:7b-instruct-q8_0 size: 7723M par: 7.6B fam: qwen2 parameter_size: 7.6B quantization_level: Q8_0 template: parameters: random check: D2 FF C6 24 8A 9B 5B 99 34 CA updating model: qwen2:7b-instruct-q8_0 pulling manifest