Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
25 changes: 25 additions & 0 deletions .github/ISSUE_TEMPLATE/bug-report..md
Original file line number Diff line number Diff line change
@@ -0,0 +1,25 @@
---
name: Bug report
about: Create a report to help us improve
title: ''
labels: bug
assignees: ''

---

## Guidelines
Please note that GitHub issues are only meant for bug reports/feature requests.


Before creating a new issue, please check whether someone else has raised the same issue. You may be able to add context to that issue instead of duplicating the report. However, each issue should also only be focussed on a _single_ problem, so do not describe new problems within an existing thread - these are very hard to track and manage, and your problem may be ignored. Finally, do not append comments to closed issues; if the same problem re-occurs, open a new issue, and include a link to the old one.

To help us understand your issue, please specify important details, primarily:

- LLM Graph builder version: X.Y.Z
- Neo4j Database version: X.Y.Z (Community/Enterprise/Aura).

- **Steps to reproduce**
- Expected behavior
- Actual behavior

Additionally, include (as appropriate) screenshots, drawings, etc.
25 changes: 25 additions & 0 deletions .github/ISSUE_TEMPLATE/feature-request.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,25 @@
---
name: Feature request
about: Suggest an idea for this project
title: ''
labels: feature
assignees: ''

---

## Guidelines
Please note that GitHub issues are only meant for bug reports/feature requests.

## Feature request template

**Is your feature request related to a problem? Please describe.**
A clear and concise description of what the problem is. Ex. I'm always frustrated when [...]

**Describe the solution you'd like**
A clear and concise description of what you want to happen.

**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.

**Additional context**
Add any other context or screenshots about the feature request here.
13 changes: 13 additions & 0 deletions .github/dependabot.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,13 @@
version: 2
updates:
- package-ecosystem: 'npm'
directory: '/frontend'
schedule:
interval: 'weekly'
target-branch: 'dev'

- package-ecosystem: 'pip'
directory: '/backend'
schedule:
interval: 'weekly'
target-branch: 'dev'
32 changes: 16 additions & 16 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -36,10 +36,6 @@ According to enviornment we are configuring the models which is indicated by VIT
EX:
```env
VITE_LLM_MODELS_PROD="openai_gpt_4o,openai_gpt_4o_mini,diffbot,gemini_1.5_flash"

You can then run Docker Compose to build and start all components:
```bash
docker-compose up --build
```

#### Additional configs
Expand Down Expand Up @@ -159,14 +155,14 @@ Allow unauthenticated request : Yes
## LLMs Supported
1. OpenAI
2. Gemini
3. Azure OpenAI(dev)
4. Anthropic(dev)
5. Fireworks(dev)
6. Groq(dev)
7. Amazon Bedrock(dev)
8. Ollama(dev)
3. Azure OpenAI(dev deployed version)
4. Anthropic(dev deployed version)
5. Fireworks(dev deployed version)
6. Groq(dev deployed version)
7. Amazon Bedrock(dev deployed version)
8. Ollama(dev deployed version)
9. Diffbot
10. Other OpenAI compabtile baseurl models(dev)
10. Other OpenAI compabtile baseurl models(dev deployed version)

## For local llms (Ollama)
1. Pull the docker imgage of ollama
Expand All @@ -177,23 +173,27 @@ docker pull ollama/ollama
```bash
docker run -d -v ollama:/root/.ollama -p 11434:11434 --name ollama ollama/ollama
```
3. Execute any llm model ex🦙3
3. Pull specific ollama model.
```bash
ollama pull llama3
```
4. Execute any llm model ex🦙3
```bash
docker exec -it ollama ollama run llama3
```
4. Configure env variable in docker compose or backend environment.
5. Configure env variable in docker compose.
```env
LLM_MODEL_CONFIG_ollama_<model_name>
#example
LLM_MODEL_CONFIG_ollama_llama3=${LLM_MODEL_CONFIG_ollama_llama3-llama3,
http://host.docker.internal:11434}
```
5. Configure the backend API url
6. Configure the backend API url
```env
VITE_BACKEND_API_URL=${VITE_BACKEND_API_URL-backendurl}
```
6. Open the application in browser and select the ollama model for the extraction.
7. Enjoy Graph Building.
7. Open the application in browser and select the ollama model for the extraction.
8. Enjoy Graph Building.


## Usage
Expand Down
6 changes: 5 additions & 1 deletion backend/example.env
Original file line number Diff line number Diff line change
Expand Up @@ -31,8 +31,10 @@ DEFAULT_DIFFBOT_CHAT_MODEL="openai_gpt_4o" #whichever model specified here , ne
LLM_MODEL_CONFIG_openai_gpt_3.5="gpt-3.5-turbo-0125,openai_api_key"
LLM_MODEL_CONFIG_openai_gpt_4o_mini="gpt-4o-mini-2024-07-18,openai_api_key"
LLM_MODEL_CONFIG_openai_gpt_4o="gpt-4o-2024-11-20,openai_api_key"
LLM_MODEL_CONFIG_openai-gpt-o3-mini="o3-mini-2025-01-31,openai_api_key"
LLM_MODEL_CONFIG_gemini_1.5_pro="gemini-1.5-pro-002"
LLM_MODEL_CONFIG_gemini_1.5_flash="gemini-1.5-flash-002"
LLM_MODEL_CONFIG_gemini_2.0_flash="gemini-2.0-flash-001"
LLM_MODEL_CONFIG_diffbot="diffbot,diffbot_api_key"
LLM_MODEL_CONFIG_azure_ai_gpt_35="azure_deployment_name,azure_endpoint or base_url,azure_api_key,api_version"
LLM_MODEL_CONFIG_azure_ai_gpt_4o="gpt-4o,https://YOUR-ENDPOINT.openai.azure.com/,azure_api_key,api_version"
Expand All @@ -44,8 +46,10 @@ LLM_MODEL_CONFIG_ollama_llama3="model_name,model_local_url"
YOUTUBE_TRANSCRIPT_PROXY="https://user:pass@domain:port"
EFFECTIVE_SEARCH_RATIO=5
GRAPH_CLEANUP_MODEL="openai_gpt_4o"
CHUNKS_TO_BE_CREATED="50"
BEDROCK_EMBEDDING_MODEL="model_name,aws_access_key,aws_secret_key,region_name" #model_name="amazon.titan-embed-text-v1"
LLM_MODEL_CONFIG_bedrock_nova_micro_v1="model_name,aws_access_key,aws_secret_key,region_name" #model_name="amazon.nova-micro-v1:0"
LLM_MODEL_CONFIG_bedrock_nova_lite_v1="model_name,aws_access_key,aws_secret_key,region_name" #model_name="amazon.nova-lite-v1:0"
LLM_MODEL_CONFIG_bedrock_nova_pro_v1="model_name,aws_access_key,aws_secret_key,region_name" #model_name="amazon.nova-pro-v1:0"
LLM_MODEL_CONFIG_fireworks_deepseek_r1="model_name,fireworks_api_key" #model_name="accounts/fireworks/models/deepseek-r1"
LLM_MODEL_CONFIG_fireworks_deepseek_v3="model_name,fireworks_api_key" #model_name="accounts/fireworks/models/deepseek-v3"
MAX_TOKEN_CHUNK_SIZE=2000 #Max token used to process/extract the file content.
39 changes: 19 additions & 20 deletions backend/requirements.txt
Original file line number Diff line number Diff line change
@@ -1,33 +1,33 @@
asyncio==3.4.3
boto3==1.35.90
botocore==1.35.90
boto3==1.36.2
botocore==1.36.2
certifi==2024.8.30
fastapi==0.115.6
fastapi-health==0.4.0
google-api-core==2.23.0
google-auth==2.36.0
google-api-core==2.24.0
google-auth==2.37.0
google_auth_oauthlib==1.2.1
google-cloud-core==2.4.1
json-repair==0.30.2
pip-install==1.3.5
langchain==0.3.13
langchain-aws==0.2.10
langchain-anthropic==0.3.0
langchain-fireworks==0.2.5
langchain-community==0.3.13
langchain-core==0.3.28
langchain==0.3.15
langchain-aws==0.2.11
langchain-anthropic==0.3.3
langchain-fireworks==0.2.6
langchain-community==0.3.15
langchain-core==0.3.31
langchain-experimental==0.3.4
langchain-google-vertexai==2.0.7
langchain-groq==0.2.1
langchain-openai==0.2.14
langchain-text-splitters==0.3.4
langchain-google-vertexai==2.0.11
langchain-groq==0.2.3
langchain-openai==0.3.1
langchain-text-splitters==0.3.5
langchain-huggingface==0.1.2
langdetect==1.0.9
langsmith==0.2.4
langserve==0.3.0
langsmith==0.2.11
langserve==0.3.1
neo4j-rust-ext
nltk==3.9.1
openai==1.58.1
openai==1.59.9
opencv-python==4.10.0.84
psutil==6.1.0
pydantic==2.9.2
Expand Down Expand Up @@ -56,7 +56,6 @@ google-cloud-logging==3.11.3
pypandoc==1.13
graphdatascience==1.12
Secweb==1.11.0
ragas==0.2.6
ragas==0.2.11
rouge_score==0.1.2
langchain-neo4j==0.2.0

langchain-neo4j==0.3.0
Loading