You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: en/ai/local-llm.md
+3-3Lines changed: 3 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -3,8 +3,8 @@
3
3
Notice:
4
4
5
5
1. This tutorial is intended for expert users
6
-
2. (Local) LLMs requires a lot of computational power
7
-
3. Smaller models (in terms of parameter size) typically respond qualitatively worse than bigger ones, but they are faster and need less memory.
6
+
2. (Local) LLMs require a lot of computational power
7
+
3. Smaller models (in terms of parameter size) typically respond qualitatively worse than bigger ones, but they are faster, need less memory and might already be sufficient for your use case.
8
8
9
9
## High-level explanation
10
10
@@ -42,5 +42,5 @@ The following steps guide you on how to use `GPT4All`to download and run local L
42
42
3. Open JabRef, go to "File" > "Preferences" > "AI"
43
43
4. Set the "AI provider" to "GPT4All"
44
44
5. Set the "Chat model" to the name (including the `.gguf`part) of the model you have downloaded in GPT4All.
45
-
6. Set the "API base URL" in "Expert Settings" to \`http://localhost:4891/v1/chat/completions\`.
45
+
6. Set the "API base URL" in "Expert Settings" to `http://localhost:4891/v1/chat/completions`.
Copy file name to clipboardExpand all lines: en/ai/preferences.md
+21-21Lines changed: 21 additions & 21 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -4,17 +4,17 @@
4
4
5
5
## General settings
6
6
7
-
- "Enable AI functionality in JabRef": by default it is turned off, so you need to check this option if you want to use the new AI features
8
-
- "Automatically generate embeddings for new entries": when this check box is switched on, for every new entry in the library, JabRef will automatically start an embeddings generation task. (If you do not know what are the embeddings, take a look at ["How does the AI functionality work?"](https://docs.jabref.org/ai#how-does-the-ai-functionality-work)).
9
-
- "Automatically generate summaries for new entries": when this check box is switched on, for every new entry in the library, JabRef will automatically generate a summary.
7
+
* "Enable AI functionality in JabRef": by default it is turned off, so you need to check this option if you want to use the new AI features
8
+
* "Automatically generate embeddings for new entries": when this check box is switched on, for every new entry in the library, JabRef will automatically start an embeddings generation task. (If you do not know what are the embeddings, take a look at ["How does the AI functionality work?"](https://docs.jabref.org/ai#how-does-the-ai-functionality-work)).
9
+
* "Automatically generate summaries for new entries": when this check box is switched on, for every new entry in the library, JabRef will automatically generate a summary.
10
10
11
11
If you import a lot of entries at a time, we recommend you to switch off options "Automatically generate embeddings for new entries" and "Automatically generate summaries for new entries", because this may slow down your computer, and you may reach the usage limit of the AI provider.
12
12
13
13
## Connection settings
14
14
15
-
- "AI provider": you can choose either OpenAI, Mistral AI, or Hugging Face
16
-
- "Chat model": choose the model you like (for OpenAI we recommend `gpt-4o-mini`, as to date, it is the cheapest and fastest, though we also recommend to look up the prices periodically, as they are subject to change)
17
-
- "API token": enter your API token here
15
+
* "AI provider": you can choose between [various providers](https://docs.jabref.org/ai/ai-providers-and-api-keys#what-is-an-ai-provider).
16
+
* "Chat model": choose the model you like.
17
+
* "API key": enter your API key here.
18
18
19
19
## Expert settings
20
20
@@ -34,7 +34,7 @@ You do not have to set this parameter manually or remember all the addresses. Ja
34
34
35
35
The embedding model transforms a document (or a piece of text) into a vector (an ordered collection of numbers). This transformation provides the AI with relevant information for your questions.
36
36
37
-
Different embedding models have varying performance, including accuracy and the speed of computing embeddings. The `_q` at the end of the model name usually denotes *quantized* (meaning reduced or simplified). These models are faster and smaller than their original counterparts but provide slightly less accuracy.
37
+
Different embedding models have varying performance, including accuracy and the speed of computing embeddings. The `_q` at the end of the model name usually denotes _quantized_ (meaning reduced or simplified). These models are faster and smaller than their original counterparts but provide slightly less accuracy.
38
38
39
39
Currently, only local embedding models are supported. This means you do not need to provide a new API key, as all the processing will be done on your machine.
40
40
@@ -112,27 +112,27 @@ To use the templates, we employ the [Apache Velocity](https://velocity.apache.or
112
112
113
113
There are four templates that JabRef uses:
114
114
115
-
-**System Message for Chatting**: This template constructs the system message (also known as the instruction) for every AI chat in JabRef (whether chatting with an entry or with a group).
116
-
-**User Message for Chatting**: This template is also used in chats and is responsible for forming a request to AI with document embeddings. The user message created by this template is sent to AI; however, only the plain user question will be saved in the chat history.
117
-
-**Summarization Chunk**: In cases where the chat model does not have enough context window to fit the entire document in one message, our algorithm will split the document into chunks. This template is used to summarize a single chunk of a document.
118
-
-**Summarization Combine**: This template is used only when the document size exceeds the context window of a chat model. It combines the summarized chunks into one piece of text.
115
+
***System Message for Chatting**: This template constructs the system message (also known as the instruction) for every AI chat in JabRef (whether chatting with an entry or with a group).
116
+
***User Message for Chatting**: This template is also used in chats and is responsible for forming a request to AI with document embeddings. The user message created by this template is sent to AI; however, only the plain user question will be saved in the chat history.
117
+
***Summarization Chunk**: In cases where the chat model does not have enough context window to fit the entire document in one message, our algorithm will split the document into chunks. This template is used to summarize a single chunk of a document.
118
+
***Summarization Combine**: This template is used only when the document size exceeds the context window of a chat model. It combines the summarized chunks into one piece of text.
119
119
120
120
You can create any template you want, but we advise starting from the default template, as it has been carefully designed and includes special syntax from Apache Velocity.
121
121
122
122
### Contexts for Templates
123
123
124
124
For each template, there is a context that holds all necessary variables used in the template. In this section, we will show you the available variables for each template and their structure.
125
125
126
-
-**System Message for Chatting**: There is a single variable, `entries`, which is a list of BIB entries. You can use `CanonicalBibEntry.getCanonicalRepresentation(BibEntry entry)` to format the entries.
127
-
-**User Message for Chatting**: There are two variables: `message` (the user question) and `excerpts` (pieces of information found in documents through the embeddings search). Each object in `excerpts` is of type `PaperExcerpt`, which has two fields: `citationKey` and `text`.
128
-
-**Summarization Chunk**: There is only the `text` variable, which contains the chunk.
129
-
-**Summarization Combine**: There is only the `chunks` variable, which contains a list of summarized chunks.
126
+
***System Message for Chatting**: There is a single variable, `entries`, which is a list of BIB entries. You can use `CanonicalBibEntry.getCanonicalRepresentation(BibEntry entry)` to format the entries.
127
+
***User Message for Chatting**: There are two variables: `message` (the user question) and `excerpts` (pieces of information found in documents through the embeddings search). Each object in `excerpts` is of type `PaperExcerpt`, which has two fields: `citationKey` and `text`.
128
+
***Summarization Chunk**: There is only the `text` variable, which contains the chunk.
129
+
***Summarization Combine**: There is only the `chunks` variable, which contains a list of summarized chunks.
130
130
131
131
## Further literature
132
132
133
-
-[Visual representation of samplers (Temperature, Top-P, Min-P, ...) by Artefact2](https://artefact2.github.io/llm-sampling/index.xhtml)
134
-
-[What is a Context Window?](https://www.techtarget.com/whatis/definition/context-window)
135
-
-[Is temperature the creativity of Large Language Models?](https://arxiv.org/abs/2405.00492)
136
-
-[The Effect of Sampling Temperature on Problem Solving in Large Language Models](https://arxiv.org/abs/2402.05201)
137
-
-[Min P Sampling: Balancing Creativity and Coherence at High Temperature](https://arxiv.org/abs/2407.01082)
138
-
-[Challenges in Deploying Long-Context Transformers: A Theoretical Peak Performance Analysis](https://arxiv.org/abs/2405.08944)
133
+
*[Visual representation of samplers (Temperature, Top-P, Min-P, ...) by Artefact2](https://artefact2.github.io/llm-sampling/index.xhtml)
134
+
*[What is a Context Window?](https://www.techtarget.com/whatis/definition/context-window)
135
+
*[Is temperature the creativity of Large Language Models?](https://arxiv.org/abs/2405.00492)
136
+
*[The Effect of Sampling Temperature on Problem Solving in Large Language Models](https://arxiv.org/abs/2402.05201)
137
+
*[Min P Sampling: Balancing Creativity and Coherence at High Temperature](https://arxiv.org/abs/2407.01082)
138
+
*[Challenges in Deploying Long-Context Transformers: A Theoretical Peak Performance Analysis](https://arxiv.org/abs/2405.08944)
Copy file name to clipboardExpand all lines: en/collect/newentryfromplaintext.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -39,7 +39,7 @@ O. Kopp, A. Armbruster, und O. Zimmermann, "Markdown Architectural Decision Reco
39
39
40
40
41
41
42
-
<figure><picture><sourcesrcset="../.gitbook/assets/Bild 5 - rule based result is selected in entry table - dark mode.png"media="(prefers-color-scheme: dark)"><imgsrc="../.gitbook/assets/new-entry-from-plain-text-step-4 (1).png"alt=""></picture><figcaption></figcaption></figure>
42
+
<figure><picture><sourcesrcset="../.gitbook/assets/Bild 5 - rule based result is selected in entry table - dark mode.png"media="(prefers-color-scheme: dark)"><imgsrc="../.gitbook/assets/Bild 5 - rule based result is selected in entry table - light mode.png"alt=""></picture><figcaption></figcaption></figure>
43
43
44
44
### Parser Explanation
45
45
@@ -51,7 +51,7 @@ This is the default parser. It does not require any extensive setups, nor does i
51
51
52
52
JabRef uses the technology offered by [Grobid](https://github.com/kermitt2/grobid), a machine learning software project with decades of experience and development dedicated to bibliographic metadata extraction. The Grobid parser usually tends to achieve better results than the rule-based parser. Since JabRef runs Grobid on a remote instance, users will have to confirm sending data to JabRef's online service in the preferences (_File > Preferences > Web search > Remote Services_). Sending data is disabled by default. It cannot be guaranteed that JabRef's Grobid instance will always be up and running, but it is possible for you to set up your [own Grobid Instance](https://grobid.readthedocs.io/en/latest/Grobid-docker/).
53
53
54
-
<figure><picture><sourcesrcset="../.gitbook/assets/Bild 6 - Grobid Preferences - dark mode.png"media="(prefers-color-scheme: dark)"><imgsrc="../.gitbook/assets/Bild 6 - Grobid Preferences - dark mode.png"alt=""></picture><figcaption><p>Grobid related preference section in JabRef</p></figcaption></figure>
54
+
<figure><picture><sourcesrcset="../.gitbook/assets/Bild 6 - Grobid Preferences - dark mode.png"media="(prefers-color-scheme: dark)"><imgsrc="../.gitbook/assets/Bild 6 - Grobid Preferences - light mode.png"alt=""></picture><figcaption><p>Grobid related preference section in JabRef</p></figcaption></figure>
0 commit comments