LLM
Scaling Llama 2 (7 - 70B) Fine-tuning on Multi-Node GPUs with Ray on Databricks
Scaling up fine-tuning and batch inferencing of LLMs such as Llama 2 (including 7B, 13B, and 70B variants) across multiple nodes without having to worry about the complexity of distributed systems.
machine-learningllama2databricks
Super Quick: Retrieval Augmented Generation Using Ollama
Unlocking the Power of Ollama Infrastructure for Local Execution of Open Source Models and Interacting with PDFs
ollamallama2llm
Structured Data from LLMs — Langchain and Pydantic Output Parser
openaillmchatgpt
The People You Need at Your Company for LLM Capabilities
artificial-intelligencellm
Your Local LLM using FastAPI
FastAPI is a modern, fast, and easy-to-use web framework for building APIs with Python. It is based on the standard Python pointer type and supports features such as data validation, documentation…
apipythonllm
Using Chains and Agents for LLM Application Development
Step-by-step guide to using chains and agents in LangChain
artificial-intelligencellmllm-models
Fine-Tuning Llama 2.0 with Single GPU Magic
Efficient-Tuning your own Language Model
artificial-intelligencellmllama-2
How to Run the LLAMA Web UI on Collab or Locally?
artificial-intelligencellmllama-2
Llama 2: A New LLMs Family Has Arrived
Use Llama 2 LLMs with Hugging Face and Transformers
artificial-intelligencellmllama-2
How to Use Llama 2 with an API on AWS to Power Your AI Apps
artificial-intelligencellmllama-2