{ "runId": "run_26f2ad91-e396-4be9-8e74-ef8a790be638", "bundleId": "llamacpp-lfm2-1.2b-q4_0.gguf-3a510a", "status": "verified", "promptTokens": 40960, "completionTokens": 10240, "contextLength": 5120, "harness": { "version": "0.1.21", "gitSha": "753787f" }, "runtime": { "name": "llama.cpp", "version": "b8640", "buildFlags": "metal" }, "model": { "displayName": "LFM2-1.2B", "format": "gguf", "quant": "q4_0", "architecture": "lfm2", "source": "LiquidAI/LFM2-1.2B-GGUF:LFM2-1.2B-Q4_0.gguf", "fileSizeBytes": 695749568, "lab": { "name": "Liquid AI", "slug": "liquid" }, "quantizedBy": { "name": "Liquid AI", "slug": "liquid" } }, "device": { "cpu": "Apple M1 Max", "cpuCores": 10, "gpu": "Apple M1 Max", "gpuCores": 32, "gpuCount": 1, "ramGb": 64, "osName": "macOS", "osVersion": "15.6.1" }, "decodeTpsMean": 178.9, "prefillTpsMean": 2146.1, "ttftP50Ms": 2311.77, "idleTpsMean": 975.2, "peakRssMb": 1204.5, "trialsPassed": 10, "trialsTotal": 10, "runnabilityScore": 0.9457756951032366, "bundleSha256": "8144bee29e019828d614d623c9552160710fbb374d3e9db869ed7944f38e2222", "createdAt": "2026-04-14T15:45:18.094Z"}