{ "runId": "run_ccfc39fa-cc75-432e-85e5-22766acda9d6", "bundleId": "llamacpp-lfm2-1.2b-extract-f16.gguf-ac3860", "status": "verified", "promptTokens": 40960, "completionTokens": 10240, "contextLength": 5120, "harness": { "version": "0.1.21", "gitSha": "4af3357" }, "runtime": { "name": "llama.cpp", "version": "b1", "buildFlags": "metal" }, "model": { "displayName": "LFM2-1.2B-Extract", "format": "gguf", "quant": "f16", "architecture": "lfm2", "source": "LiquidAI/LFM2-1.2B-Extract-GGUF:LFM2-1.2B-Extract-F16.gguf", "fileSizeBytes": 2343325408, "lab": { "name": "Liquid AI", "slug": "liquid" }, "quantizedBy": { "name": "Liquid AI", "slug": "liquid" } }, "device": { "cpu": "Intel(R) Core(TM) i7-5930K CPU @ 3.50GHz", "cpuCores": 12, "gpu": "NVIDIA TITAN Xp", "gpuCores": 0, "gpuCount": 2, "ramGb": 31, "osName": "Ubuntu 24.04.4 LTS", "osVersion": "24.04" }, "decodeTpsMean": 153.8, "prefillTpsMean": 4088.9, "ttftP50Ms": 1001.9, "idleTpsMean": 712.8, "peakRssMb": 2583.7, "trialsPassed": 10, "trialsTotal": 10, "runnabilityScore": 0.9651178175403226, "bundleSha256": "4a3e95739021d7aadf94d0dab1fe2db1ac310ae37e13b645140670cc59702d15", "createdAt": "2026-04-14T02:14:42.308Z"}