mirror of
https://github.com/danswer-ai/danswer.git
synced 2025-03-26 17:51:54 +01:00
Remove wordnet (#2365)
This commit is contained in:
parent
1555ac9dab
commit
148c2a7375
@ -75,8 +75,8 @@ Tokenizer.from_pretrained('nomic-ai/nomic-embed-text-v1')"
|
||||
# Pre-downloading NLTK for setups with limited egress
|
||||
RUN python -c "import nltk; \
|
||||
nltk.download('stopwords', quiet=True); \
|
||||
nltk.download('wordnet', quiet=True); \
|
||||
nltk.download('punkt', quiet=True);"
|
||||
# nltk.download('wordnet', quiet=True); introduce this back if lemmatization is needed
|
||||
|
||||
# Set up application files
|
||||
WORKDIR /app
|
||||
|
@ -3,7 +3,6 @@ from collections.abc import Callable
|
||||
|
||||
import nltk # type:ignore
|
||||
from nltk.corpus import stopwords # type:ignore
|
||||
from nltk.stem import WordNetLemmatizer # type:ignore
|
||||
from nltk.tokenize import word_tokenize # type:ignore
|
||||
from sqlalchemy.orm import Session
|
||||
|
||||
@ -40,7 +39,7 @@ logger = setup_logger()
|
||||
def download_nltk_data() -> None:
|
||||
resources = {
|
||||
"stopwords": "corpora/stopwords",
|
||||
"wordnet": "corpora/wordnet",
|
||||
# "wordnet": "corpora/wordnet", # Not in use
|
||||
"punkt": "tokenizers/punkt",
|
||||
}
|
||||
|
||||
@ -58,15 +57,16 @@ def download_nltk_data() -> None:
|
||||
|
||||
|
||||
def lemmatize_text(keywords: list[str]) -> list[str]:
|
||||
try:
|
||||
query = " ".join(keywords)
|
||||
lemmatizer = WordNetLemmatizer()
|
||||
word_tokens = word_tokenize(query)
|
||||
lemmatized_words = [lemmatizer.lemmatize(word) for word in word_tokens]
|
||||
combined_keywords = list(set(keywords + lemmatized_words))
|
||||
return combined_keywords
|
||||
except Exception:
|
||||
return keywords
|
||||
raise NotImplementedError("Lemmatization should not be used currently")
|
||||
# try:
|
||||
# query = " ".join(keywords)
|
||||
# lemmatizer = WordNetLemmatizer()
|
||||
# word_tokens = word_tokenize(query)
|
||||
# lemmatized_words = [lemmatizer.lemmatize(word) for word in word_tokens]
|
||||
# combined_keywords = list(set(keywords + lemmatized_words))
|
||||
# return combined_keywords
|
||||
# except Exception:
|
||||
# return keywords
|
||||
|
||||
|
||||
def remove_stop_words_and_punctuation(keywords: list[str]) -> list[str]:
|
||||
|
Loading…
x
Reference in New Issue
Block a user