|
1 | 1 | --- |
2 | 2 | layout: base |
3 | | -title: "Graph of text data, using large language models" |
4 | | -keywords: "LLMs, Retrieval Augmented Generation" |
| 3 | +title: "Exploring the AI brain using graphs" |
| 4 | +keywords: "LLMs, latent space, Retrieval Augmented Generation" |
5 | 5 | contact_person: "Benjamin Ricaud" |
6 | 6 | --- |
7 | 7 |
|
8 | 8 |
|
9 | 9 | ## 📝 Description |
10 | | -The main idea is to build graphs of documents, graphs of topics contained in them, connecting abstract concepts and ideas inside the documents. We want to detect important news, automatically classify documents and get detailed information about a topic. The approach is in the same vein as Retrieval Augmented Generation [RAG](https://arxiv.org/abs/2005.11401) but we want to leverage the graph structure for a better retrieval of information. |
| 10 | +The main goal of this project is a better understanding of how large language models structure their inner representation of data, how their "AI brain" is organized. We will collect the embeddings of a AI model for a dataset of documents. The idea is then to create the graphs of embeddings and topics contained in them. By encoding chuncks of text following each other in documents, we want to see the connection between abstract concepts and ideas from the documents inside the AI model. |
| 11 | +We will explore the representation of sentences and texts of (small) text encoders such as BERT and [modernBERT](https://arxiv.org/abs/2412.13663) its recent updated version. To map the latent space we will use an approach similar to Retrieval-Augmented Generation [RAG](https://arxiv.org/abs/2005.11401) and create an graph of nearest neighbors in the latent space to structure it. |
| 12 | +We will first explore the structure of the graph, see if there are clusters, central concepts or ideas inside the model representation. Then we will then use the graph to evaluate how the AI model chain or connect ideas in documents and as it produces text. |
| 13 | +In this project we will make use of simple and advanced tools from network analysis as well as machine learning and generative AI (LLMs). We will use pre-trained models from Huggingface. |
11 | 14 |
|
12 | | -Graphs will be made using the inner representation of a large laguage model. we will start by feeding a deep encoder model such as ada002 or BERT with documents or sections of documents and build a vector database from the embeddings of the model. We will then make a graph of nearest neighbors. We will first explore the structure of the graph, see if there are clusters, central concepts or documents. We will then use the graph to refine the feedback given for a given prompt to the LLM. |
13 | | -In this project we will make use of simple and advanced tools from network analysis as well as machine learning and generative AI (LLMs). |
14 | | - |
15 | | -**Data:** Open datasets of documents, databases form the Norwegian public services, Wikipedia data |
| 15 | +**Data:** Open datasets of documents, Wikipedia data |
16 | 16 |
|
17 | 17 | ## 📨 Contact: |
18 | | -Benjamin Ricaud < [email protected]> |
| 18 | +Benjamin Ricaud < [email protected]> |
0 commit comments