You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+4-2Lines changed: 4 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,17 +1,19 @@
1
1
---
2
2
title: Large language model topic modelling
3
-
emoji: 📝
3
+
emoji: 📚
4
4
colorFrom: purple
5
5
colorTo: yellow
6
6
sdk: gradio
7
+
sdk_version: 6.0.2
7
8
app_file: app.py
8
9
pinned: true
9
10
license: agpl-3.0
11
+
short_description: Create thematic summaries for open text data
10
12
---
11
13
12
14
# Large language model topic modelling
13
15
14
-
Version: 0.5.1
16
+
Version: 0.5.2
15
17
16
18
Extract topics and summarise outputs using Large Language Models (LLMs, Gemma 3 4b/GPT-OSS 20b if local (see tools/config.py to modify), Gemini, Azure, or AWS Bedrock models (e.g. Claude, Nova models). The app will query the LLM with batches of responses to produce summary tables, which are then compared iteratively to output a table with the general topics, subtopics, topic sentiment, and a topic summary. Instructions on use can be found in the README.md file. You can try out examples by clicking on one of the example datasets on the main app page, which will show you example outputs from a local model run. API keys for AWS, Azure, and Gemini services can be entered on the settings page (note that Gemini has a free public API).
description = "Generate thematic summaries from open text in tabular data files with a large language model."
9
9
requires-python = ">=3.10"
10
10
readme = "README.md"
@@ -50,55 +50,59 @@ classifiers = [
50
50
"Programming Language :: Python :: 3.13",
51
51
]
52
52
53
-
dependencies = [
54
-
"pandas==2.3.3",
55
-
"gradio==5.49.1",
56
-
"transformers==4.57.1",
53
+
dependencies = [
54
+
"gradio==6.0.2",
55
+
"transformers==4.57.2",
57
56
"spaces==0.42.1",
58
-
"boto3==1.40.72",
59
-
"pyarrow==21.0.0",
60
-
"openpyxl==3.1.5",
61
-
"markdown==3.7",
62
-
"tabulate==0.9.0",
63
-
"lxml==5.3.0",
64
-
"google-genai==1.50.0",
65
-
"openai==2.2.0",
66
-
"html5lib==1.1",
67
-
"beautifulsoup4==4.12.3",
68
-
"rapidfuzz==3.13.0",
69
-
"python-dotenv==1.1.0"
57
+
"boto3==1.42.1",
58
+
"pandas<=2.3.3",
59
+
"pyarrow>=21.0.0",
60
+
"openpyxl>=3.1.5",
61
+
"markdown>=3.7",
62
+
"tabulate>=0.9.0",
63
+
"lxml>=5.3.0",
64
+
"google-genai<=1.52.0",
65
+
"openai<=2.8.1",
66
+
"html5lib>=1.1",
67
+
"beautifulsoup4>=4.12.3",
68
+
"rapidfuzz>=3.13.0",
69
+
"python-dotenv>=1.1.0"
70
70
]
71
71
72
72
[project.optional-dependencies]
73
73
dev = ["pytest"]
74
74
test = ["pytest", "pytest-cov"]
75
75
76
76
# Extra dependencies for VLM models
77
-
# For torch you should use --index-url https://download.pytorch.org/whl/cu124. Additionally installs the unsloth package
77
+
# For torch you should use --index-url https://download.pytorch.org/whl/cu128. Additionally installs the unsloth package
78
78
torch = [
79
-
"torch==2.6.0",
80
-
"accelerate==1.11.0",
81
-
"bitsandbytes==0.48.2",
82
-
"unsloth==2025.9.4",
83
-
"unsloth_zoo==2025.9.5",
84
-
"timm==1.0.19"
79
+
"torch<=2.9.1",
80
+
"torchvision",
81
+
"accelerate",
82
+
"bitsandbytes",
83
+
"unsloth==2025.11.6",
84
+
"unsloth_zoo==2025.11.6",
85
+
"timm",
86
+
"xformers"
85
87
]
86
88
87
-
# If you want to install llama-cpp-python in GPU mode, use cmake.args="-DGGML_CUDA=on" . If that doesn't work, try specific wheels for your system, e.g. for Linux: See files in https://github.com/abetlen/llama-cpp-python/releases/tag/v0.3.16-cu124 . More details on installation here: https://llama-cpp-python.readthedocs.io/en/latest
89
+
# If you want to install llama-cpp-python in GPU mode, use cmake.args="-DGGML_CUDA=on" . If that doesn't work, try specific wheels for your system, e.g. for Linux see files in https://github.com/JamePeng/llama-cpp-python/releases. More details on installation here: https://llama-cpp-python.readthedocs.io/en/latest
# Note that this requirements file is optimised for Hugging Face spaces / Python 3.10. Please use requirements_no_local.txt for installation without local model inference (simplest approach to get going). Please use requirements_cpu.txt for CPU instances and requirements_gpu.txt for GPU instances using Python 3.11
# If above doesn't work, try specific wheels for your system, see files in https://github.com/JamePeng/llama-cpp-python/releases for different python versions
# This requirements file is optimised for AWS ECS using Python 3.11 alongside the Dockerfile, without local torch and llama-cpp-python. For AWS ECS, torch and llama-cpp-python are optionally installed in the main Dockerfile
0 commit comments