llm
A list of posts tagged llm
Blogs
- Configure Ollama on Dev Containers and VS Code
- Getting started with Ollama on Windows
- Using Generative AI to produce Spotify Clips
Notes
- Clock Tables - Org Mode, Plain Text, and AI
- These models are too damn big!
- What about instrumentals? AI Generated Spotify Clips Addendum
- AI abundance after scarcity cycles
Responses
- Bringing Llama 3 to life
- GPT-4o System Card
- LongROPE: Extending LLM Context Window Beyond 2 Million Tokens
- Home-Cooked Software and Barefoot Developers
- Meta Large Language Model Compiler: Foundation Models of Compiler Optimization
- Mapping the Mind of a Large Language Model
- Ultravox - An open, fast, and extensible multimodal LLM
- Reproducing GPT-2 (124M) in llm.c in 90 minutes for $20
- Introducing Snowflake Arctic
- Introducing Phi-3
- Introducing Llama 3
- Griffin: Mixing Gated Linear Recurrences with Local Attention for Efficient Language Models
- DE-COP: Detecting Copyrighted Content in Language Models Training Data
- Using LLM to select the right SQL Query from candidates
- RecurrentGemma - Open weights language model from Google DeepMind, based on Griffin.
- Hello OLMo: A truly open LLM
- LLM training in simple, raw C/CUDA
- ARAGOG: Advanced RAG Output Grading
- OpenAI - Introducing improvements to the fine-tuning API and expanding our custom models program
- Introducing Command R+: A Scalable LLM Built for Business
- Large Language Models Are Zero-Shot Time Series Forecasters
- IPEX-LLM
- Start using ChatGPT instantly
- Announcing DBRX: A new standard for efficient open source LLMs
- Mamba: Linear-Time Sequence Modeling with Selective State Spaces
- Releasing Common Corpus: the largest public domain dataset for training LLMs
- Demystifying Embedding Spaces using Large Language Models
- MM1: Methods, Analysis & Insights from Multimodal LLM Pre-training
- Ollama now supports AMD graphics cards
- You can now train a 70b language model at home
- Levels of Complexity: RAG Applications
- Inflection-2.5: meet the world's best personal AI
- Training great LLMs entirely from ground up in the wilderness as a startup
- Gemma PyTorch
- Introducing the next generation of Claude
- GGUF, the long way around
- Predictive Human Preference: From Model Ranking to Model Routing
- The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits
- Announcing Mistral Large
- GPT in 500 lines of SQL
- Gemma: Introducing new state-of-the-art open models
- Cosmopedia v0.1
- Ollama - Windows Preview
- GraphRAG: Unlocking LLM discovery on narrative private data
- NVIDIA Chat with RTX
- Memory and new controls for ChatGPT
- Eagle 7B : Soaring past Transformers with 1 Trillion Tokens Across 100+ Languages (RWKV-v5)
- Google’s Hugging Face deal puts ‘supercomputer’ power behind open-source AI
- New OpenAI embedding models and API updates
- NightShade
- Introducing Stable LM 2 1.6B
- Stable Code 3B - Coding on the Edge
- Sampling for Text Generation
- Talking about Open Source LLMs on Oxide and Friends
- LeftoverLocals: Listening to LLM responses through leaked GPU local memory
- AI for economists - prompts and resources
- More than an OpenAI Wrapper: Perplexity Pivots to Open Source
- My AI Timelines Have Sped Up (Again)
- Ferret: Refer and Ground Anything Anywhere at Any Granularity
- LangChain State of AI 2023
- LLM in a flash: Efficient Large Language Model Inference with Limited Memory
- OpenAI - Prompt engineering
- Bash One-Liners for LLMs
- Phi-2: The surprising power of small language models
- Introducing Stable LM Zephyr 3B
- The Geometry of Truth: Dataexplorer
- MemGPT - Towards LLMs as Operating Systems
- Best Practices for LLM Evaluation of RAG Applications
- MLflow 2.8 with LLM-as-a-judge metrics and Best Practices for LLM Evaluation of RAG Applications
- OpenAgents: An Open Platform for Language Agents in the Wild
- Mixtral of experts
- Long context prompting for Claude 2.1
- Introducing llamafile
- Chain-of-Verification Reduces Hallucination in Large Language Models
- Scaffolded LLMs as natural language computers
- Mistral 7B Model
- Distilling step-by-step: Outperforming larger language models with less training data and smaller model sizes
- vim + llm = 🔥
- ChatGPT can now see, hear, and speak
- DALL·E 3
- Optimizing your LLM in production
- PointLLM: Empowering Large Language Models to Understand Point Clouds
- Spread Your Wings: Falcon 180B is here
- Generative AI and .NET - Part 2 SDK
- Perplexity: Interactive LLM visualization
- Can LLMs learn from a single example?
- Generative AI and .NET - Part 1 Intro
- Llama 2 7B/13B are now available in Web LLM
- Supporting the Open Source AI Community
- Making Large Language Models Work For You
- Introducing Code Llama, a state-of-the-art large language model for coding
- Large Language Models with Semantic Search
- Patterns for Building LLM-based Systems and Products
- Open challenges in LLM research
- GPT-4 API general availability
- Gorilla: Large Language Model Connected with Massive APIs
- LoRA: Low-Rank Adaptation of Large Language Models
- Language models can explain neurons in language models
- Free Dolly
- Generative Agents: Interactive Simulacra of Human Behavior
- Koala: A Dialogue Model for Academic Research