Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
8 changes: 4 additions & 4 deletions src/embedders/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@
from spacy.tokens.doc import Doc
from sklearn.decomposition import PCA
from tqdm import tqdm
from embedders import util
from src.embedders import util
from joblib import dump, load


Expand Down Expand Up @@ -112,7 +112,7 @@ def __init__(
embedder: Embedder,
n_components: int = 8,
autocorrect_n_components: bool = True,
**kwargs
**kwargs,
):
super().__init__()
self.embedder = embedder
Expand All @@ -126,15 +126,15 @@ def store_pca_weights(self, file_name: str):
Args:
file_name (str): Path to the file without any file endings.
"""
dump(self.reducer, f'{file_name}.joblib')
dump(self.reducer, f"{file_name}.joblib")

def load_pca_weights(self, file_name: str):
"""Loads the PCA weights from a file.

Args:
file_name (str): Path to the file without any file endings.
"""
self.reducer = load(f'{file_name}.joblib')
self.reducer = load(f"{file_name}.joblib")

@abstractmethod
def _reduce(
Expand Down
2 changes: 1 addition & 1 deletion src/embedders/classification/__init__.py
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
from embedders import Embedder
from src.embedders import Embedder


class SentenceEmbedder(Embedder):
Expand Down
4 changes: 2 additions & 2 deletions src/embedders/classification/contextual.py
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
from typing import List, Optional, Union, Generator
from sentence_transformers import SentenceTransformer
from embedders import util
from embedders.classification import SentenceEmbedder
from src.embedders import util
from src.embedders.classification import SentenceEmbedder
from spacy.tokens.doc import Doc
import torch
import openai
Expand Down
4 changes: 2 additions & 2 deletions src/embedders/classification/count_based.py
Original file line number Diff line number Diff line change
@@ -1,8 +1,8 @@
from typing import List, Union, Generator
from sklearn.feature_extraction.text import CountVectorizer
from sklearn.feature_extraction.text import TfidfVectorizer
from embedders.classification import SentenceEmbedder
from embedders import util
from src.embedders.classification import SentenceEmbedder
from src.embedders import util


class CountSentenceEmbedder(SentenceEmbedder):
Expand Down
2 changes: 1 addition & 1 deletion src/embedders/classification/reduce.py
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
from spacy.tokens.doc import Doc
from typing import Union, List, Generator
import numpy as np
from embedders import PCAReducer, util
from src.embedders import PCAReducer, util


class PCASentenceReducer(PCAReducer):
Expand Down
2 changes: 1 addition & 1 deletion src/embedders/extraction/__init__.py
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
import spacy
from embedders import Embedder
from src.embedders import Embedder
from spacy.tokens.doc import Doc
from typing import Union

Expand Down
6 changes: 3 additions & 3 deletions src/embedders/extraction/contextual.py
Original file line number Diff line number Diff line change
Expand Up @@ -5,12 +5,12 @@
import re
from transformers import AutoTokenizer, AutoModel
from collections import defaultdict
from embedders import util
from src.embedders import util
from spacy.tokens.doc import Doc


from embedders.enums import WarningType
from embedders.extraction import TokenEmbedder
from src.embedders.enums import WarningType
from src.embedders.extraction import TokenEmbedder


class TransformerTokenEmbedder(TokenEmbedder):
Expand Down
4 changes: 2 additions & 2 deletions src/embedders/extraction/count_based.py
Original file line number Diff line number Diff line change
@@ -1,9 +1,9 @@
from typing import List, Generator, Union
from sklearn.feature_extraction.text import CountVectorizer
from embedders import util
from src.embedders import util
from spacy.tokens.doc import Doc

from embedders.extraction import TokenEmbedder
from src.embedders.extraction import TokenEmbedder


class BagOfCharsTokenEmbedder(TokenEmbedder):
Expand Down
2 changes: 1 addition & 1 deletion src/embedders/extraction/reduce.py
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
from typing import List, Generator, Union
import numpy as np
from embedders import PCAReducer, util
from src.embedders import PCAReducer, util


class PCATokenReducer(PCAReducer):
Expand Down