🧪 Skills
Openclaw Skill Langchain Local
Run LangChain AI pipelines locally and offline using Ollama and phi4-mini for coding, devops, chat, or document-based queries without requiring cloud or API...
v0.1.0
Description
LangChain Local — Offline AI Chains with Ollama
Run LangChain pipelines locally using Ollama and phi4-mini (or any local model). No API keys, no cloud, fully private and offline.
Requirements
- Ollama installed and running (
ollama serve) - phi4-mini pulled (
ollama pull phi4-mini) - Python packages:
pip install langchain langchain-ollama langchain-community
Modes
coding— Python/Django code generation (low temperature, precise)devops— Linux/Nginx/Docker shell commandschat— General conversationrag— Document-grounded answers (context-aware)
Usage Examples
- "Ask langchain-local to write a Django REST API view"
- "Use langchain-local in devops mode to show how to check disk usage"
- "Run langchain-local chat mode: explain what is RAG"
- "Use langchain-local coding mode to write a PostgreSQL backup script"
Implementation
Execute: python3 ~/.openclaw/skills/langchain-local/main.py
Parameters
- mode: coding | devops | chat | rag (default: coding)
- query: Your question or instruction
Author
Manoj — https://github.com/manoj
Reviews (0)
Sign in to write a review.
No reviews yet. Be the first to review!
Comments (0)
No comments yet. Be the first to share your thoughts!