Chatbot:Run flowise and ollama locally

Prerequisites

Server:

ollama run llama3.2 
ollama run nomic-embed-text:la

Install Flowise locally using NPM:

npm install -g flowise

Start Flowise:

npx flowise start

if successful:

Steps:

1 . Create documents stores

2. Click Document store and select document loader (eg: pdf file)

  1. Preview and process

  1. Embedding using nomic-embed-text as we run from ollama library

5.Vector store

Select faiss vector store

  1. Run upsert and if successful as below

7 . The last step is start chat flow and add nodes. Now we have finished build chatbot using ollama 3.2 locally