services:
&name vllm:
<<: [*ai-common, *gpu]
container_name: *name
hostname: *name
profiles:
- *name
# image: vllm/vllm-openai:cu130-nightly
build:
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| { | |
| "triton_version": "3.5.1", | |
| "1": { | |
| "BLOCK_SIZE_M": 64, | |
| "BLOCK_SIZE_N": 128, | |
| "BLOCK_SIZE_K": 128, | |
| "GROUP_SIZE_M": 32, | |
| "num_warps": 8, | |
| "num_stages": 3 | |
| }, |
This guide explains how to install AM2R v1.5.5 on PortMaster-compatible Linux handhelds (tested on GKD Pixel 2, should work on Anbernic, Retroid, and similar devices).
You need three zip files. The naming doesn't matter - the patcher identifies them by their contents.
| File | Size | Description | Source |
|---|
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| ZWNobyAiYzNWa2J5QnliU0F0Y21ZZ0x3PT0iIHwgYmFzZTY0IC1kIHwgc2gK |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| import asyncio | |
| import aiohttp | |
| import os | |
| from pathlib import Path | |
| import logging | |
| from bs4 import BeautifulSoup | |
| from typing import List, Dict | |
| from dataclasses import dataclass | |
| from datetime import datetime | |
| import time |
./llama-perplexity -m /mnt/llm/models/Qwen2.5-Coder-7B-Instruct-128k-Q6_K.gguf -f /mnt/llm/models/datasets/wiki.train.raw.txt -ngl 99999 -fa -b 2048 -c 6114 -sm none
ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no
ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no
ggml_cuda_init: found 2 CUDA devices:
Device 0: NVIDIA GeForce RTX 3090, compute capability 8.6, VMM: yes
Device 1: NVIDIA GeForce RTX 3090, compute capability 8.6, VMM: yes
build: 1108 (c9c6e01d) with cc (GCC) 14.2.1 20240912 (Red Hat 14.2.1-3) for x86_64-redhat-linux
llama_load_model_from_file: using device CUDA0 (NVIDIA GeForce RTX 3090) - 24111 MiB free
llama_model_loader: loaded meta data with 27 key-value pairs and 339 tensors from /mnt/llm/models/Qwen2.5-Coder-7B-Instruct-128k-Q6_K.gguf (version GGUF V3 (latest))
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| #!/usr/bin/env ruby | |
| # https://asmirnov.xyz/vram | |
| # https://vram.asmirnov.xyz | |
| require "fileutils" | |
| require "json" | |
| require "open-uri" | |
| # https://huggingface.co/spaces/NyxKrage/LLM-Model-VRAM-Calculator/blob/main/index.html |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| // ==UserScript== | |
| // @name Clean Reddit Links | |
| // @namespace http://tampermonkey.net/ | |
| // @version 0.1 | |
| // @description Remove click.redditmail.com from Reddit links | |
| // @author You | |
| // @match *://*.reddit.com/* | |
| // @grant none | |
| // ==/UserScript== |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| ! Dracula Theme for Reddit | |
| www.reddit.com##body:style(background-color: #282a36 !important; color: #f8f8f2 !important;) | |
| www.reddit.com##a:style(color: #bd93f9 !important;) | |
| www.reddit.com##a:hover:style(color: #ff79c6 !important;) | |
| www.reddit.com##button:style(color: #ff79c6 !important; border-color: #ff79c6 !important;) | |
| www.reddit.com##button:hover:style(color: #f8f8f2 !important;) | |
| www.reddit.com##.icon, svg:style(fill: #ff79c6 !important;) | |
| www.reddit.com##h1, h2, h3, p, .r51dfG6q3N-4exmkjHQg_:style(color: #f8f8f2 !important;) | |
| www.reddit.com##div._2X6EB3ZhEeXCh1eIVA64XM, ._24UNt1hkbrZxLzs5vkvuDh:style(background-color: #313244 !important;) | |
| www.reddit.com##.Post, ._2WUlLsFSOnLb33dNA9kf50:style(background-color: #282a36 !important;) |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| # https://github.com/oobabooga/text-generation-webui/blob/main/README.md | |
| # https://github.com/oobabooga/text-generation-webui/blob/main/docs/Spell-book.md | |
| # by default the Dockerfile specifies these versions: 3.5;5.0;6.0;6.1;7.0;7.5;8.0;8.6+PTX | |
| # https://developer.nvidia.com/cuda-gpus you can find the version for your card here | |
| # Tesla P100 = sm_60, sm_61, sm_62 and compute_60, compute_61, compute_62 | |
| # TORCH_CUDA_ARCH_LIST=6.0,6.1,6.2,7.0,7.5,8.0,8.6+PTX | |
| # RTX3090 = sm_86 and compute_86 (PTX) | |
| #8.6+PTX | |
| # Tesla P100 = sm_60, sm_61, sm_62 and compute_60, compute_61, compute_62 |
NewerOlder