|
| 1 | +# Created with komac v2.15.0 |
| 2 | +# yaml-language-server: $schema=https://aka.ms/winget-manifest.defaultLocale.1.12.0.schema.json |
| 3 | + |
| 4 | +PackageIdentifier: ggml.llamacpp |
| 5 | +PackageVersion: b8833 |
| 6 | +PackageLocale: en-US |
| 7 | +Publisher: ggml |
| 8 | +PublisherUrl: https://github.com/ggml-org |
| 9 | +PublisherSupportUrl: https://github.com/ggml-org/llama.cpp/issues |
| 10 | +PackageName: llama.cpp |
| 11 | +PackageUrl: https://github.com/ggml-org/llama.cpp |
| 12 | +License: MIT |
| 13 | +LicenseUrl: https://github.com/ggml-org/llama.cpp/blob/HEAD/LICENSE |
| 14 | +ShortDescription: LLM inference in C/C++ |
| 15 | +Tags: |
| 16 | +- ggml |
| 17 | +- llama |
| 18 | +ReleaseNotes: |- |
| 19 | + ggml-webgpu: fix compiler warnings and refactor FlashAttention encoding (#21052) |
| 20 | + - Update workflows to remove dependence on llvmpipe |
| 21 | + - Try setting Dawn_DIR |
| 22 | + - remove c++20 initializers |
| 23 | + - Move to proper guid |
| 24 | + - Try avoiding segfaults on vulkan backend process exit |
| 25 | + - Remove compiler warnings on parameter casting |
| 26 | + - Fix soft_max and update reg_tile accumulation to f32 for better precision |
| 27 | + - Refactor flash_attn a bit |
| 28 | + - remove c++20 initializers and format |
| 29 | + - Increase div precision for NVIDIA |
| 30 | + - revert div precision and comment out ggml-ci node for now |
| 31 | + - Formatting |
| 32 | + - Try debugging on a failing CI node |
| 33 | + - Revert "Try debugging on a failing CI node" |
| 34 | + This reverts commit 1971e33. |
| 35 | + macOS/iOS: |
| 36 | + - macOS Apple Silicon (arm64) |
| 37 | + - macOS Apple Silicon (arm64, KleidiAI enabled) |
| 38 | + - macOS Intel (x64) |
| 39 | + - iOS XCFramework |
| 40 | + Linux: |
| 41 | + - Ubuntu x64 (CPU) |
| 42 | + - Ubuntu arm64 (CPU) |
| 43 | + - Ubuntu s390x (CPU) |
| 44 | + - Ubuntu x64 (Vulkan) |
| 45 | + - Ubuntu arm64 (Vulkan) |
| 46 | + - Ubuntu x64 (ROCm 7.2) |
| 47 | + - Ubuntu x64 (OpenVINO) |
| 48 | + Android: |
| 49 | + - Android arm64 (CPU) |
| 50 | + Windows: |
| 51 | + - Windows x64 (CPU) |
| 52 | + - Windows arm64 (CPU) |
| 53 | + - Windows x64 (CUDA 12) - CUDA 12.4 DLLs |
| 54 | + - Windows x64 (CUDA 13) - CUDA 13.1 DLLs |
| 55 | + - Windows x64 (Vulkan) |
| 56 | + - Windows x64 (SYCL) |
| 57 | + - Windows x64 (HIP) |
| 58 | + openEuler: |
| 59 | + - openEuler x86 (310p) |
| 60 | + - openEuler x86 (910b, ACL Graph) |
| 61 | + - openEuler aarch64 (310p) |
| 62 | + - openEuler aarch64 (910b, ACL Graph) |
| 63 | +ReleaseNotesUrl: https://github.com/ggml-org/llama.cpp/releases/tag/b8833 |
| 64 | +Documentations: |
| 65 | +- DocumentLabel: Wiki |
| 66 | + DocumentUrl: https://github.com/ggml-org/llama.cpp/wiki |
| 67 | +ManifestType: defaultLocale |
| 68 | +ManifestVersion: 1.12.0 |
0 commit comments