why reddit?
Phind V7 subjectively performing at GPT4 levels for coding
Min P sampler (an alternative to Top K/Top P) has been merged into llama.cpp
HUGE dataset released for open source use
I've started uploading quants of exllama v2 models, taking requests
Nearly 10% of people ask AI chatbots for explicit content. Will it lead LLMs astray?
Text Generation Web-UI has been updated to CUDA 12.1, and with it new docker images are needed
Single Digit tokenization improves LLM math abilities by up to 70x
Musical notation
Are Local LLMs Useful in Incident Response? - SANS Internet Storm Center
Dolphin 2.0 based on mistral-7b released by Eric Hartford
Beginner questions thread
Mistral 7B model
Microsoft's latest LLM agent: autogen
QA-LoRA: Quantization-Aware Low-Rank Adaptation of Large Language Models
Effective Long-Context Scaling of Foundation Models | Research - AI at Meta
Jeremy Howard: A Hackers' Guide to Language Models
Amazon investing in Anthropic - Expanding access to safer AI with Amazon
Very interesting thread about reversal knowledge
Draft & Verify: Lossless Large Language Model Acceleration via Self-Speculative Decoding