Techietalks AI - Multi-Model Chat Assistant

Welcome to the Techietalks AI repository! This project is a multi-model AI-powered chat assistant built using Streamlit, Pydantic_AI, and multiple AI models like OpenAI, DeepSeek, and Gemini. It allows users to interact with different AI models, switch between them, and enjoy a conversational experience. The app also includes a "New Chat" button to reset the conversation and start fresh.

This repository contains all the necessary files to set up and run the chat assistant locally or in a Docker container. Below, you'll find a detailed explanation of the project and instructions to get started.


πŸ“ Repository Structure

Here’s a breakdown of the files in this repository:

  1. .DS_Store: A macOS-specific file that stores folder attributes (e.g., icon positions). You can ignore this file.
  2. .env: A file to store environment variables, such as API keys. Important: Add your OPENAI_API_KEY, DEEPSEEK_API_KEY, and GEMINI_API_KEY here to authenticate with the respective APIs.
  3. .gitignore: Specifies files and folders that Git should ignore (e.g., .env to avoid exposing sensitive information).
  4. Dockerfile: Contains instructions to build a Docker image for the application. It sets up a Python environment, installs dependencies, and runs the Streamlit app.
  5. README.md: This file! It provides an overview of the project and instructions for setup.
  6. app.py: The main application file. It contains the code for the Streamlit-based chat interface and integrates with multiple AI models (OpenAI, DeepSeek, and Gemini). It also includes a "New Chat" button to reset the conversation.
  7. docker-compose.yml: A configuration file to run the application using Docker Compose. It maps ports, loads environment variables, and sets up volumes.
  8. requirements.txt: Lists all Python dependencies required to run the application.
  9. sree.txt: A placeholder text file (likely for personal notes or testing).

πŸš€ How It Works

This project uses the following technologies:

When you type a question into the chat interface, the app sends it to the selected AI model (OpenAI, DeepSeek, or Gemini), processes the response, and displays it in a conversational format. It also maintains a history of the conversation, allowing the AI to provide more context-aware answers. The "New Chat" button resets the conversation, clearing the history and starting fresh.


πŸ› οΈ Setup Instructions

Follow these steps to set up and run the project:

Prerequisites

  1. Python 3.11: Ensure Python is installed on your system.
  2. Docker (optional): If you want to run the app in a container.
  3. API Keys: Obtain API keys for OpenAI, DeepSeek, and Gemini from their respective websites.

Steps

  1. Clone the Repository: bash git clone https://github.com/schogini/techietalksai.git cd techietalksai/0005

  2. Set Up Environment Variables:

  3. Install Dependencies:

  4. Run the Application:


πŸ“„ License

This project is open-source and available under the MIT License. Feel free to use, modify, and distribute it as needed.


πŸ“§ Contact

For AI consultancy, training, and development, contact Schogini Systems Private Limited at https://www.schogini.com.


πŸ‘¨β€πŸ’» Author

Sreeprakash Neelakantan
- Website: https://www.schogini.com
- GitHub: https://github.com/schogini/techietalksai.git


Enjoy using the Techietalks AI Multi-Model Chat Assistant! If you have any questions or feedback, feel free to reach out. 😊