auto download vader and punkt_tab instead of making the user download them automatically
This commit is contained in:
parent
be85c9f30d
commit
f74b310b47
2 changed files with 26 additions and 9 deletions
1
.gitignore
vendored
1
.gitignore
vendored
|
@ -5,3 +5,4 @@ MEMORY_LOADED
|
||||||
memory.json
|
memory.json
|
||||||
*.pkl
|
*.pkl
|
||||||
*.json
|
*.json
|
||||||
|
venv/
|
32
bot.py
32
bot.py
|
@ -3,8 +3,6 @@ from discord.ext import commands, tasks
|
||||||
from discord import app_commands
|
from discord import app_commands
|
||||||
import json
|
import json
|
||||||
import markovify
|
import markovify
|
||||||
import nltk
|
|
||||||
from nltk.tokenize import word_tokenize
|
|
||||||
import random
|
import random
|
||||||
import os
|
import os
|
||||||
import time
|
import time
|
||||||
|
@ -20,15 +18,35 @@ from better_profanity import profanity
|
||||||
from config import *
|
from config import *
|
||||||
import traceback
|
import traceback
|
||||||
import shutil
|
import shutil
|
||||||
from nltk.sentiment.vader import SentimentIntensityAnalyzer
|
import nltk
|
||||||
|
from nltk.data import find
|
||||||
|
from nltk import download
|
||||||
|
|
||||||
analyzer = SentimentIntensityAnalyzer()
|
|
||||||
|
|
||||||
print(splashtext) # you can use https://patorjk.com/software/taag/ for 3d text or just remove this entirely
|
print(splashtext) # you can use https://patorjk.com/software/taag/ for 3d text or just remove this entirely
|
||||||
|
|
||||||
|
def check_resources():
|
||||||
|
resources = {
|
||||||
|
'vader_lexicon': 'sentiment/vader_lexicon',
|
||||||
|
'punkt_tab': 'tokenizers/punkt',
|
||||||
|
}
|
||||||
|
|
||||||
|
for resource, path in resources.items():
|
||||||
|
try:
|
||||||
|
find(path)
|
||||||
|
print(f"{resource} is already installed.")
|
||||||
|
except LookupError:
|
||||||
|
print(f"{resource} is not installed. Downloading now...")
|
||||||
|
download(resource)
|
||||||
|
|
||||||
|
|
||||||
|
from nltk.sentiment.vader import SentimentIntensityAnalyzer
|
||||||
|
from nltk.tokenize import word_tokenize
|
||||||
|
analyzer = SentimentIntensityAnalyzer()
|
||||||
|
|
||||||
def download_json():
|
def download_json():
|
||||||
locales_dir = "locales"
|
locales_dir = "locales"
|
||||||
response = requests.get(f"{VERSION_URL}/goob/locales/{LOCALE}.json")
|
response = requests.get(f"https://raw.githubusercontent.com/gooberinc/goober/refs/heads/main/locales/{LOCALE}.json")
|
||||||
if response.status_code == 200:
|
if response.status_code == 200:
|
||||||
|
|
||||||
if not os.path.exists(locales_dir):
|
if not os.path.exists(locales_dir):
|
||||||
|
@ -42,7 +60,7 @@ def download_json():
|
||||||
|
|
||||||
if not os.path.exists(os.path.join(locales_dir, "en.json")):
|
if not os.path.exists(os.path.join(locales_dir, "en.json")):
|
||||||
|
|
||||||
response = requests.get(f"{VERSION_URL}/goob/locales/en.json")
|
response = requests.get(f"https://raw.githubusercontent.com/gooberinc/goober/refs/heads/main/locales/en.json")
|
||||||
if response.status_code == 200:
|
if response.status_code == 200:
|
||||||
with open(os.path.join(locales_dir, "en.json"), "w", encoding="utf-8") as file:
|
with open(os.path.join(locales_dir, "en.json"), "w", encoding="utf-8") as file:
|
||||||
file.write(response.text)
|
file.write(response.text)
|
||||||
|
@ -223,8 +241,6 @@ def get_file_info(file_path):
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
return {"error": str(e)}
|
return {"error": str(e)}
|
||||||
|
|
||||||
nltk.download('punkt')
|
|
||||||
|
|
||||||
def load_memory():
|
def load_memory():
|
||||||
data = []
|
data = []
|
||||||
|
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue