Answers your AI related questions with information from research papers using a RAG LLM system
Nvidia GPU
Docker
Docker-compose
https://www.docker.com/get-started/
https://docs.docker.com/compose/install/
run command
git clone https://github.com/segallagher/Citegeist.git
- Create
.envfile with the following values \
DATASET_PATH="data/arxiv_dataset/dataset.csv"
VECTORSTORE_DIR="vectorstore"
PAPER_DIR="published/papers"- Assemble papers into one csv file by running
python .\dataset_creation\assemble_dataset.py - Create vectorstore from dataset
python .\create_vectorstore_db.py
Enter directory for Citegeist in terminal
cd Citegeist
run docker-compse
docker-compose up
In your browser go to the webui at
http://localhost:80
In the chatbox, enter your question and get your answers.
Add the following to your .env file \
EMBED_MODEL_TYPE="ollama"
EMBED_MODEL="mxbai-embed-large:latest"
LLM_MODEL_TYPE="ollama"
LLM_MODEL="llama3.2:3b"
OLLAMA_HOST="http://localhost:11434"- Evaluation happens in
evaluate_rag.py. Depending on which type of evaluation you want to do you must set theOPERATIONparameter to the type of evaluation you want to do. - The questions are stored in
data_analysis/questions.json
Each question will be answered by the RAG LLM system and then judged byllama3.1:8b - Run
evaluate_rag.py
This will likely take a while running ollama on your computer