This repository showcases my hands-on projects while learning about AI Agents using LangChain, Ollama, and Python. It includes both minimal setups and advanced interactive applications that demonstrate real-world agent use cases like memory handling, voice interaction, and local LLM inference.
. ├── basic_ai_agent/ # Basic text agent with and without memory │ ├── basic_ai_agent.py │ ├── basic_ai_agent_with_memory.py │ ├── streamlit_app.py │ ├── requirements.txt │ └── README.md │ ├── pdf_summary_bot/ # QA-based PDF summarizer agent │ ├── app_basic_qa.py │ ├── app_summary_qa.py │ ├── requirements.txt │ └── README.md │ ├── voice_assistant/ # Voice assistant with terminal + Streamlit UI │ ├── ai_voice_assistant_and_streamlit_app.py │ ├── streamlit_app.py │ ├── requirements.txt │ └── README.md │ ├── web_scraper_agent/ # Agent that scrapes data using tools + memory │ ├── ai_web_scraper.py │ ├── ai_web_scrapper_faiss.py │ ├── requirements.txt │ └── README.md │ ├── .gitignore ├── README.md # ← You are here └── requirements.txt # Master requirements for full repo
| Feature | Library/Tool |
|---|---|
| LLM Inference | Ollama |
| Language Agent APIs | LangChain, LangChain Community |
| Web UI | Streamlit |
| Voice I/O | SpeechRecognition, pyttsx3 |
| Parsing & Memory | FAISS, LangChain Memory |
Minimal examples to:
- Load local LLMs (like
phi,mistral, etc.) via Ollama - Run single-turn and memory-supported agents
- Try a web UI version with Streamlit
Ask questions or summarize based on uploaded PDF content:
app_basic_qa.py: Basic question answeringapp_summary_qa.py: Summary-focused version
Talk to your LLM with:
- A terminal-based CLI voice chat
- A full Streamlit-based voice chat UI
Give queries like "Find top Python tutorials" and get results scraped live:
ai_web_scraper.py: Basic tool-based agentai_web_scrapper_faiss.py: Uses memory and FAISS for smarter recall
-
Clone this repo
git clone https://github.com/your-username/ai-agent-projects.git cd ai-agent-projects -
Install Dependencies
pip install -r requirements.txt
-
Make sure Ollama is running
ollama run phi
-
Run a project
cd voice_assistant python ai_voice_assistant_and_streamlit_app.py # OR streamlit run streamlit_app.py
Each folder includes its own requirements.txt, but a master file is available at root if you want to install everything together.
- Python 3.9+
- Ollama running locally with a model like
phi,mistral, orllama2 - Microphone + speaker (for voice apps)
- LangChain for the powerful agent ecosystem
- Ollama for local, private LLM serving
- Streamlit for interactive UIs
This repo is for educational purposes. MIT License. Feel free to fork, use, and modify with attribution.
- Add file upload / PDF reading support
- Build a full research assistant
- Experiment with tools, memory, and agents
Pull requests and suggestions are welcome!