mirror of
https://github.com/Alexander-D-Karpov/akarpov
synced 2024-11-24 02:03:49 +03:00
Compare commits
1 Commits
909e0deb38
...
a71aa8beec
Author | SHA1 | Date | |
---|---|---|---|
|
a71aa8beec |
|
@ -15,5 +15,3 @@ LAST_FM_SECRET=
|
||||||
SPOTIFY_ID=
|
SPOTIFY_ID=
|
||||||
SPOTIFY_SECRET=
|
SPOTIFY_SECRET=
|
||||||
YANDEX_TOKEN=
|
YANDEX_TOKEN=
|
||||||
PREVIEW_SERVICE_API_KEY=
|
|
||||||
PREVIEW_SERVICE_URL=
|
|
||||||
|
|
15
.github/workflows/ci.yml
vendored
15
.github/workflows/ci.yml
vendored
|
@ -24,10 +24,10 @@ jobs:
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- name: Checkout Code Repository
|
- name: Checkout Code Repository
|
||||||
uses: actions/checkout@v4.2.1
|
uses: actions/checkout@v3
|
||||||
|
|
||||||
- name: Cache packages
|
- name: Cache packages
|
||||||
uses: actions/cache@v4.1.1
|
uses: actions/cache@v3
|
||||||
id: cache-packages
|
id: cache-packages
|
||||||
with:
|
with:
|
||||||
path: "~/packages/"
|
path: "~/packages/"
|
||||||
|
@ -45,18 +45,18 @@ jobs:
|
||||||
sudo dpkg -L libimage-exiftool-perl libmagickwand-dev | while IFS= read -r f; do if test -f $f; then echo $f; fi; done | xargs cp --parents --target-directory ~/packages/
|
sudo dpkg -L libimage-exiftool-perl libmagickwand-dev | while IFS= read -r f; do if test -f $f; then echo $f; fi; done | xargs cp --parents --target-directory ~/packages/
|
||||||
fi
|
fi
|
||||||
|
|
||||||
- uses: actions/checkout@v4.2.1
|
- uses: actions/checkout@v3
|
||||||
- name: Install poetry
|
- name: Install poetry
|
||||||
run: pipx install poetry
|
run: pipx install poetry
|
||||||
|
|
||||||
- uses: actions/setup-python@v5.2.0
|
- uses: actions/setup-python@v5
|
||||||
with:
|
with:
|
||||||
python-version: '3.11'
|
python-version: '3.11'
|
||||||
cache: 'poetry'
|
cache: 'poetry'
|
||||||
- run: poetry install
|
- run: poetry install
|
||||||
|
|
||||||
- name: Run pre-commit
|
- name: Run pre-commit
|
||||||
uses: pre-commit/action@v3.0.1
|
uses: pre-commit/action@v2.0.3
|
||||||
|
|
||||||
# With no caching at all the entire ci process takes 4m 30s to complete!
|
# With no caching at all the entire ci process takes 4m 30s to complete!
|
||||||
pytest:
|
pytest:
|
||||||
|
@ -64,10 +64,7 @@ jobs:
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- name: Checkout Code Repository
|
- name: Checkout Code Repository
|
||||||
uses: actions/checkout@v4.2.1
|
uses: actions/checkout@v3
|
||||||
|
|
||||||
- name: Install Docker Compose
|
|
||||||
run: sudo apt-get update && sudo apt-get install -y docker-compose
|
|
||||||
|
|
||||||
- name: Build the Stack
|
- name: Build the Stack
|
||||||
run: docker-compose -f local.yml build
|
run: docker-compose -f local.yml build
|
||||||
|
|
|
@ -28,6 +28,10 @@ $ uvicorn redirect.app:app --reload
|
||||||
```shell
|
```shell
|
||||||
$ docker-compose -f local.yml up
|
$ docker-compose -f local.yml up
|
||||||
```
|
```
|
||||||
|
Install file preview dependencies
|
||||||
|
```shell
|
||||||
|
$ docker-compose -f local.yml exec django /install_preview_dependencies
|
||||||
|
```
|
||||||
- server - http://127.0.0.1:8000
|
- server - http://127.0.0.1:8000
|
||||||
- mail - http://127.0.0.1:8025
|
- mail - http://127.0.0.1:8025
|
||||||
|
|
||||||
|
|
0
akarpov/common/ml/__init__.py
Normal file
0
akarpov/common/ml/__init__.py
Normal file
51
akarpov/common/ml/text.py
Normal file
51
akarpov/common/ml/text.py
Normal file
|
@ -0,0 +1,51 @@
|
||||||
|
import pycld2 as cld2
|
||||||
|
import spacy
|
||||||
|
import torch
|
||||||
|
from transformers import AutoModel, AutoTokenizer
|
||||||
|
|
||||||
|
# load ml classes and models on first request
|
||||||
|
# TODO: move to outer server/service
|
||||||
|
nlp = None
|
||||||
|
ru_nlp = None
|
||||||
|
|
||||||
|
ru_model = None
|
||||||
|
ru_tokenizer = None
|
||||||
|
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
||||||
|
|
||||||
|
|
||||||
|
def get_text_embedding(text: str):
|
||||||
|
global nlp, ru_nlp, ru_model, ru_tokenizer
|
||||||
|
|
||||||
|
is_reliable, text_bytes_found, details = cld2.detect(text)
|
||||||
|
if is_reliable:
|
||||||
|
lang = details[0]
|
||||||
|
if lang[1] in ["ru", "en"]:
|
||||||
|
lang = lang[1]
|
||||||
|
else:
|
||||||
|
return None
|
||||||
|
else:
|
||||||
|
return None
|
||||||
|
|
||||||
|
if lang == "ru":
|
||||||
|
if not ru_nlp:
|
||||||
|
ru_nlp = spacy.load("ru_core_news_md", disable=["parser", "ner"])
|
||||||
|
lema = " ".join([token.lemma_ for token in ru_nlp(text)])
|
||||||
|
if not ru_model:
|
||||||
|
ru_model = AutoModel.from_pretrained("DeepPavlov/rubert-base-cased")
|
||||||
|
if not ru_tokenizer:
|
||||||
|
ru_tokenizer = AutoTokenizer.from_pretrained("DeepPavlov/rubert-base-cased")
|
||||||
|
encodings = ru_tokenizer(
|
||||||
|
lema, # the texts to be tokenized
|
||||||
|
padding=True, # pad the texts to the maximum length (so that all outputs have the same length)
|
||||||
|
return_tensors="pt", # return the tensors (not lists)
|
||||||
|
)
|
||||||
|
with torch.no_grad():
|
||||||
|
# get the model embeddings
|
||||||
|
embeds = ru_model(**encodings)
|
||||||
|
embeds = embeds[0]
|
||||||
|
elif lang == "en":
|
||||||
|
embeds = None
|
||||||
|
else:
|
||||||
|
embeds = None
|
||||||
|
|
||||||
|
return embeds
|
|
@ -1,10 +1,16 @@
|
||||||
|
import textract
|
||||||
|
|
||||||
from akarpov.files.models import File
|
from akarpov.files.models import File
|
||||||
|
|
||||||
|
|
||||||
def view(file: File):
|
def view(file: File):
|
||||||
static = ""
|
static = ""
|
||||||
content = ""
|
content = ""
|
||||||
text = file.content.replace("\t", " ")
|
text = (
|
||||||
|
textract.process(file.file.path, extension="doc", output_encoding="utf8")
|
||||||
|
.decode("utf8")
|
||||||
|
.replace("\t", " ")
|
||||||
|
)
|
||||||
for line in text.split("\n"):
|
for line in text.split("\n"):
|
||||||
content += f"<p class='mt-1'>{line}</p>"
|
content += f"<p class='mt-1'>{line}</p>"
|
||||||
return static, content
|
return static, content
|
||||||
|
|
|
@ -1,10 +1,16 @@
|
||||||
|
import textract
|
||||||
|
|
||||||
from akarpov.files.models import File
|
from akarpov.files.models import File
|
||||||
|
|
||||||
|
|
||||||
def view(file: File):
|
def view(file: File):
|
||||||
static = ""
|
static = ""
|
||||||
content = ""
|
content = ""
|
||||||
text = file.content.replace("\t", " ")
|
text = (
|
||||||
|
textract.process(file.file.path, extension="docx", output_encoding="utf8")
|
||||||
|
.decode("utf8")
|
||||||
|
.replace("\t", " ")
|
||||||
|
)
|
||||||
for line in text.split("\n"):
|
for line in text.split("\n"):
|
||||||
content += f"<p class='mt-1'>{line}</p>"
|
content += f"<p class='mt-1'>{line}</p>"
|
||||||
return static, content
|
return static, content
|
||||||
|
|
|
@ -1,10 +1,16 @@
|
||||||
|
import textract
|
||||||
|
|
||||||
from akarpov.files.models import File
|
from akarpov.files.models import File
|
||||||
|
|
||||||
|
|
||||||
def view(file: File):
|
def view(file: File):
|
||||||
static = ""
|
static = ""
|
||||||
content = ""
|
content = ""
|
||||||
text = file.content.replace("\t", " ")
|
text = (
|
||||||
|
textract.process(file.file.path, extension="odt", output_encoding="utf8")
|
||||||
|
.decode("utf8")
|
||||||
|
.replace("\t", " ")
|
||||||
|
)
|
||||||
for line in text.split("\n"):
|
for line in text.split("\n"):
|
||||||
content += f"<p class='mt-1'>{line}</p>"
|
content += f"<p class='mt-1'>{line}</p>"
|
||||||
return static, content
|
return static, content
|
||||||
|
|
|
@ -1,3 +1,5 @@
|
||||||
|
import textract
|
||||||
|
|
||||||
from akarpov.files.models import File
|
from akarpov.files.models import File
|
||||||
|
|
||||||
|
|
||||||
|
@ -5,7 +7,11 @@ def view(file: File) -> (str, str):
|
||||||
static = f"""
|
static = f"""
|
||||||
<meta property="og:title" content="{file.name}" />
|
<meta property="og:title" content="{file.name}" />
|
||||||
"""
|
"""
|
||||||
text = file.content.replace("\t", " ")
|
text = (
|
||||||
|
textract.process(file.file.path, extension="ogg", output_encoding="utf8")
|
||||||
|
.decode("utf8")
|
||||||
|
.replace("\t", " ")
|
||||||
|
)
|
||||||
content = (
|
content = (
|
||||||
"""
|
"""
|
||||||
<div id="waveform">
|
<div id="waveform">
|
||||||
|
|
42
akarpov/files/services/lema.py
Normal file
42
akarpov/files/services/lema.py
Normal file
|
@ -0,0 +1,42 @@
|
||||||
|
from nltk.corpus import stopwords
|
||||||
|
from nltk.stem import WordNetLemmatizer
|
||||||
|
from nltk.tokenize import word_tokenize
|
||||||
|
from pymorphy3 import MorphAnalyzer
|
||||||
|
|
||||||
|
# Set up stop words
|
||||||
|
english_stopwords = set(stopwords.words("english"))
|
||||||
|
russian_stopwords = set(stopwords.words("russian"))
|
||||||
|
|
||||||
|
# Set up lemmatizers
|
||||||
|
english_lemmatizer = None
|
||||||
|
russian_lemmatizer = None
|
||||||
|
|
||||||
|
|
||||||
|
def lemmatize_and_remove_stopwords(text, language="english"):
|
||||||
|
# Tokenize the text
|
||||||
|
global english_lemmatizer, russian_lemmatizer
|
||||||
|
tokens = word_tokenize(text)
|
||||||
|
|
||||||
|
# Lemmatize each token based on the specified language
|
||||||
|
lemmatized_tokens = []
|
||||||
|
for token in tokens:
|
||||||
|
if language == "russian":
|
||||||
|
if not russian_lemmatizer:
|
||||||
|
russian_lemmatizer = MorphAnalyzer()
|
||||||
|
lemmatized_token = russian_lemmatizer.parse(token)[0].normal_form
|
||||||
|
else: # Default to English
|
||||||
|
if not english_lemmatizer:
|
||||||
|
english_lemmatizer = WordNetLemmatizer()
|
||||||
|
lemmatized_token = english_lemmatizer.lemmatize(token)
|
||||||
|
lemmatized_tokens.append(lemmatized_token)
|
||||||
|
|
||||||
|
# Remove stop words
|
||||||
|
filtered_tokens = [
|
||||||
|
token
|
||||||
|
for token in lemmatized_tokens
|
||||||
|
if token not in english_stopwords and token not in russian_stopwords
|
||||||
|
]
|
||||||
|
|
||||||
|
# Reconstruct the text
|
||||||
|
filtered_text = " ".join(filtered_tokens)
|
||||||
|
return filtered_text
|
|
@ -1,4 +1,8 @@
|
||||||
|
from math import ceil
|
||||||
|
|
||||||
import magic
|
import magic
|
||||||
|
from PIL import Image, ImageDraw, ImageFont
|
||||||
|
from preview_generator.manager import PreviewManager
|
||||||
|
|
||||||
from akarpov.files.models import File
|
from akarpov.files.models import File
|
||||||
|
|
||||||
|
@ -15,11 +19,90 @@
|
||||||
manager = None
|
manager = None
|
||||||
|
|
||||||
|
|
||||||
|
def textfile_to_image(textfile_path) -> Image:
|
||||||
|
"""Convert text file to a grayscale image.
|
||||||
|
|
||||||
|
arguments:
|
||||||
|
textfile_path - the content of this file will be converted to an image
|
||||||
|
font_path - path to a font file (for example impact.ttf)
|
||||||
|
"""
|
||||||
|
# parse the file into lines stripped of whitespace on the right side
|
||||||
|
with open(textfile_path) as f:
|
||||||
|
lines = tuple(line.rstrip() for line in f.readlines())
|
||||||
|
|
||||||
|
font: ImageFont = None
|
||||||
|
large_font = 20 # get better resolution with larger size
|
||||||
|
for font_filename in COMMON_MONO_FONT_FILENAMES:
|
||||||
|
try:
|
||||||
|
font = ImageFont.truetype(font_filename, size=large_font)
|
||||||
|
print(f'Using font "{font_filename}".')
|
||||||
|
break
|
||||||
|
except OSError:
|
||||||
|
print(f'Could not load font "{font_filename}".')
|
||||||
|
if font is None:
|
||||||
|
font = ImageFont.load_default()
|
||||||
|
print("Using default font.")
|
||||||
|
|
||||||
|
def _font_points_to_pixels(pt):
|
||||||
|
return round(pt * 96.0 / 72)
|
||||||
|
|
||||||
|
margin_pixels = 20
|
||||||
|
|
||||||
|
# height of the background image
|
||||||
|
tallest_line = max(lines, key=lambda line: font.getsize(line)[PIL_HEIGHT_INDEX])
|
||||||
|
max_line_height = _font_points_to_pixels(
|
||||||
|
font.getsize(tallest_line)[PIL_HEIGHT_INDEX]
|
||||||
|
)
|
||||||
|
realistic_line_height = max_line_height * 0.8
|
||||||
|
image_height = int(ceil(realistic_line_height * len(lines) + 2 * margin_pixels))
|
||||||
|
|
||||||
|
widest_line = max(lines, key=lambda s: font.getsize(s)[PIL_WIDTH_INDEX])
|
||||||
|
max_line_width = _font_points_to_pixels(font.getsize(widest_line)[PIL_WIDTH_INDEX])
|
||||||
|
image_width = int(ceil(max_line_width + (2 * margin_pixels)))
|
||||||
|
|
||||||
|
# draw the background
|
||||||
|
background_color = 255 # white
|
||||||
|
image = Image.new(
|
||||||
|
PIL_GRAYSCALE, (image_width, image_height), color=background_color
|
||||||
|
)
|
||||||
|
draw = ImageDraw.Draw(image)
|
||||||
|
|
||||||
|
font_color = 0
|
||||||
|
horizontal_position = margin_pixels
|
||||||
|
for i, line in enumerate(lines):
|
||||||
|
vertical_position = int(round(margin_pixels + (i * realistic_line_height)))
|
||||||
|
draw.text(
|
||||||
|
(horizontal_position, vertical_position), line, fill=font_color, font=font
|
||||||
|
)
|
||||||
|
|
||||||
|
return image
|
||||||
|
|
||||||
|
|
||||||
|
def create_preview(file_path: str) -> str:
|
||||||
|
global manager
|
||||||
|
# TODO: add text image generation/code image
|
||||||
|
if not manager:
|
||||||
|
manager = PreviewManager(cache_path, create_folder=True)
|
||||||
|
if manager.has_jpeg_preview(file_path):
|
||||||
|
return manager.get_jpeg_preview(file_path, height=500)
|
||||||
|
return ""
|
||||||
|
|
||||||
|
|
||||||
def get_file_mimetype(file_path: str) -> str:
|
def get_file_mimetype(file_path: str) -> str:
|
||||||
mime = magic.Magic(mime=True)
|
mime = magic.Magic(mime=True)
|
||||||
return mime.from_file(file_path)
|
return mime.from_file(file_path)
|
||||||
|
|
||||||
|
|
||||||
|
def get_description(file_path: str) -> str:
|
||||||
|
global manager
|
||||||
|
if not manager:
|
||||||
|
manager = PreviewManager(cache_path, create_folder=True)
|
||||||
|
|
||||||
|
if manager.has_text_preview(file_path):
|
||||||
|
return manager.get_text_preview(file_path)
|
||||||
|
return ""
|
||||||
|
|
||||||
|
|
||||||
def get_base_meta(file: File):
|
def get_base_meta(file: File):
|
||||||
preview = file.preview.url if file.preview else ""
|
preview = file.preview.url if file.preview else ""
|
||||||
description = file.description if file.description else ""
|
description = file.description if file.description else ""
|
||||||
|
|
|
@ -11,6 +11,12 @@
|
||||||
from akarpov.files.models import File
|
from akarpov.files.models import File
|
||||||
|
|
||||||
from ..documents import FileDocument
|
from ..documents import FileDocument
|
||||||
|
from .lema import lemmatize_and_remove_stopwords
|
||||||
|
|
||||||
|
"""
|
||||||
|
Calculus on types of searches:
|
||||||
|
https://new.akarpov.ru/files/FZUTFBIyfbdlDHVzxUNU
|
||||||
|
"""
|
||||||
|
|
||||||
|
|
||||||
class BaseSearch:
|
class BaseSearch:
|
||||||
|
@ -134,20 +140,23 @@ class SimilaritySearch(BaseSearch):
|
||||||
def search(self, query: str) -> QuerySet[File]:
|
def search(self, query: str) -> QuerySet[File]:
|
||||||
if self.queryset is None:
|
if self.queryset is None:
|
||||||
raise ValueError("Queryset cannot be None for similarity search")
|
raise ValueError("Queryset cannot be None for similarity search")
|
||||||
|
|
||||||
|
language = "russian" if re.search("[а-яА-Я]", query) else "english"
|
||||||
|
filtered_query = lemmatize_and_remove_stopwords(query, language=language)
|
||||||
queryset = (
|
queryset = (
|
||||||
self.queryset.annotate(
|
self.queryset.annotate(
|
||||||
name_similarity=Coalesce(
|
name_similarity=Coalesce(
|
||||||
TrigramSimilarity(UnaccentLower("name"), query),
|
TrigramSimilarity(UnaccentLower("name"), filtered_query),
|
||||||
Value(0),
|
Value(0),
|
||||||
output_field=FloatField(),
|
output_field=FloatField(),
|
||||||
),
|
),
|
||||||
description_similarity=Coalesce(
|
description_similarity=Coalesce(
|
||||||
TrigramSimilarity(UnaccentLower("description"), query),
|
TrigramSimilarity(UnaccentLower("description"), filtered_query),
|
||||||
Value(0),
|
Value(0),
|
||||||
output_field=FloatField(),
|
output_field=FloatField(),
|
||||||
),
|
),
|
||||||
content_similarity=Coalesce(
|
content_similarity=Coalesce(
|
||||||
TrigramSimilarity(UnaccentLower("content"), query),
|
TrigramSimilarity(UnaccentLower("content"), filtered_query),
|
||||||
Value(0),
|
Value(0),
|
||||||
output_field=FloatField(),
|
output_field=FloatField(),
|
||||||
),
|
),
|
||||||
|
|
18
akarpov/files/services/text.py
Normal file
18
akarpov/files/services/text.py
Normal file
|
@ -0,0 +1,18 @@
|
||||||
|
import chardet
|
||||||
|
import textract
|
||||||
|
from textract.exceptions import ExtensionNotSupported
|
||||||
|
|
||||||
|
|
||||||
|
def extract_file_text(file: str) -> str:
|
||||||
|
try:
|
||||||
|
text = textract.process(file)
|
||||||
|
except ExtensionNotSupported:
|
||||||
|
try:
|
||||||
|
rawdata = open(file, "rb").read()
|
||||||
|
enc = chardet.detect(rawdata)
|
||||||
|
with open(file, encoding=enc["encoding"]) as f:
|
||||||
|
text = f.read()
|
||||||
|
except Exception:
|
||||||
|
return ""
|
||||||
|
|
||||||
|
return text
|
|
@ -1,69 +1,40 @@
|
||||||
import base64
|
import os
|
||||||
import time
|
import time
|
||||||
from urllib.parse import urljoin
|
|
||||||
|
|
||||||
import requests
|
|
||||||
import structlog
|
import structlog
|
||||||
from celery import shared_task
|
from celery import shared_task
|
||||||
from django.conf import settings
|
|
||||||
from django.core import management
|
from django.core import management
|
||||||
from django.core.files.base import ContentFile
|
from django.core.files import File
|
||||||
|
|
||||||
from akarpov.files.models import File as FileModel
|
from akarpov.files.models import File as FileModel
|
||||||
|
from akarpov.files.services.preview import create_preview, get_file_mimetype
|
||||||
|
from akarpov.files.services.text import extract_file_text
|
||||||
|
|
||||||
logger = structlog.get_logger(__name__)
|
logger = structlog.get_logger(__name__)
|
||||||
|
|
||||||
|
|
||||||
def sanitize_content(content):
|
|
||||||
"""Remove NUL (0x00) characters from the content."""
|
|
||||||
if isinstance(content, str):
|
|
||||||
return content.replace("\x00", "")
|
|
||||||
elif isinstance(content, bytes):
|
|
||||||
return content.replace(b"\x00", b"")
|
|
||||||
return content
|
|
||||||
|
|
||||||
|
|
||||||
@shared_task()
|
@shared_task()
|
||||||
def process_file(pk: int):
|
def process_file(pk: int):
|
||||||
|
pth = None
|
||||||
file = FileModel.objects.get(pk=pk)
|
file = FileModel.objects.get(pk=pk)
|
||||||
if not file.name:
|
if not file.name:
|
||||||
file.name = file.file.name.split("/")[-1]
|
file.name = file.file.name.split("/")[-1]
|
||||||
|
|
||||||
try:
|
try:
|
||||||
api_url = urljoin(settings.PREVIEW_SERVICE_URL, "/process_file/")
|
pth = create_preview(file.file.path)
|
||||||
|
if pth:
|
||||||
files = {"file": (file.name, file.file.open("rb"))}
|
with open(pth, "rb") as f:
|
||||||
headers = {
|
file.preview.save(
|
||||||
"X-API-Key": settings.PREVIEW_SERVICE_API_KEY,
|
pth.split("/")[-1],
|
||||||
"Accept": "application/json",
|
File(f),
|
||||||
}
|
save=False,
|
||||||
|
)
|
||||||
response = requests.post(api_url, files=files, headers=headers)
|
|
||||||
|
|
||||||
if response.status_code != 200:
|
|
||||||
logger.error(f"Failed to process file {pk}: {response.text}")
|
|
||||||
return
|
|
||||||
|
|
||||||
result = response.json()
|
|
||||||
|
|
||||||
file.file_type = result["file_type"]
|
|
||||||
file.content = sanitize_content(result["content"])
|
|
||||||
|
|
||||||
if result["preview"]:
|
|
||||||
image_data = base64.b64decode(result["preview"])
|
|
||||||
file.preview.save(
|
|
||||||
f"{file.name}_preview.jpg", ContentFile(image_data), save=False
|
|
||||||
)
|
|
||||||
|
|
||||||
file.save()
|
|
||||||
|
|
||||||
logger.info(f"File {pk} processed successfully")
|
|
||||||
|
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.error(f"Error processing file {pk}: {str(e)}")
|
logger.error(e)
|
||||||
finally:
|
file.file_type = get_file_mimetype(file.file.path)
|
||||||
file.file.close()
|
file.content = extract_file_text(file.file.path)
|
||||||
|
file.save(update_fields=["preview", "name", "file_type", "content"])
|
||||||
|
if pth and os.path.isfile(pth):
|
||||||
|
os.remove(pth)
|
||||||
return pk
|
return pk
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -7,8 +7,6 @@
|
||||||
Album,
|
Album,
|
||||||
AnonMusicUser,
|
AnonMusicUser,
|
||||||
Author,
|
Author,
|
||||||
MusicDraft,
|
|
||||||
MusicDraftFile,
|
|
||||||
Playlist,
|
Playlist,
|
||||||
PlaylistSong,
|
PlaylistSong,
|
||||||
Song,
|
Song,
|
||||||
|
@ -380,61 +378,3 @@ class Meta:
|
||||||
"link": {"read_only": True},
|
"link": {"read_only": True},
|
||||||
"image": {"read_only": True},
|
"image": {"read_only": True},
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
class AllSearchSerializer(serializers.Serializer):
|
|
||||||
songs = serializers.SerializerMethodField(method_name="get_songs")
|
|
||||||
authors = serializers.SerializerMethodField(method_name="get_authors")
|
|
||||||
albums = serializers.SerializerMethodField(method_name="get_albums")
|
|
||||||
|
|
||||||
@extend_schema_field(ListSongSerializer(many=True))
|
|
||||||
def get_songs(self, obj):
|
|
||||||
return ListSongSerializer(
|
|
||||||
Song.objects.cache().search(obj["query"]).to_queryset()[:10],
|
|
||||||
many=True,
|
|
||||||
context=self.context,
|
|
||||||
).data
|
|
||||||
|
|
||||||
@extend_schema_field(ListAuthorSerializer(many=True))
|
|
||||||
def get_authors(self, obj):
|
|
||||||
return ListAuthorSerializer(
|
|
||||||
Author.objects.cache().search(obj["query"]).to_queryset()[:10], many=True
|
|
||||||
).data
|
|
||||||
|
|
||||||
@extend_schema_field(ListAlbumSerializer(many=True))
|
|
||||||
def get_albums(self, obj):
|
|
||||||
return ListAlbumSerializer(
|
|
||||||
Album.objects.cache().search(obj["query"]).to_queryset()[:10], many=True
|
|
||||||
).data
|
|
||||||
|
|
||||||
|
|
||||||
class MusicDraftFileSerializer(serializers.ModelSerializer):
|
|
||||||
class Meta:
|
|
||||||
model = MusicDraftFile
|
|
||||||
fields = ["file", "original_name", "mime_type"]
|
|
||||||
|
|
||||||
|
|
||||||
class MusicDraftSerializer(serializers.ModelSerializer):
|
|
||||||
files = MusicDraftFileSerializer(many=True, read_only=True)
|
|
||||||
|
|
||||||
class Meta:
|
|
||||||
model = MusicDraft
|
|
||||||
fields = [
|
|
||||||
"id",
|
|
||||||
"status",
|
|
||||||
"provider",
|
|
||||||
"original_url",
|
|
||||||
"meta_data",
|
|
||||||
"file_token",
|
|
||||||
"created",
|
|
||||||
"updated",
|
|
||||||
"error_message",
|
|
||||||
"files",
|
|
||||||
]
|
|
||||||
read_only_fields = ["id", "file_token", "created", "updated"]
|
|
||||||
|
|
||||||
|
|
||||||
class MusicDraftCallbackSerializer(serializers.Serializer):
|
|
||||||
status = serializers.ChoiceField(choices=MusicDraft.STATUS_CHOICES)
|
|
||||||
meta_data = serializers.JSONField(required=False)
|
|
||||||
error_message = serializers.CharField(required=False)
|
|
||||||
|
|
|
@ -16,14 +16,11 @@
|
||||||
ListSongPlaylistsAPIView,
|
ListSongPlaylistsAPIView,
|
||||||
ListSongSlugsAPIView,
|
ListSongSlugsAPIView,
|
||||||
ListUserListenedSongsAPIView,
|
ListUserListenedSongsAPIView,
|
||||||
MusicDraftCallbackView,
|
|
||||||
MusicDraftDetailView,
|
|
||||||
RemoveSongFromPlaylistAPIView,
|
RemoveSongFromPlaylistAPIView,
|
||||||
RetrieveUpdateDestroyAlbumAPIView,
|
RetrieveUpdateDestroyAlbumAPIView,
|
||||||
RetrieveUpdateDestroyAuthorAPIView,
|
RetrieveUpdateDestroyAuthorAPIView,
|
||||||
RetrieveUpdateDestroyPlaylistAPIView,
|
RetrieveUpdateDestroyPlaylistAPIView,
|
||||||
RetrieveUpdateDestroySongAPIView,
|
RetrieveUpdateDestroySongAPIView,
|
||||||
SearchAllAPIView,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
app_name = "music"
|
app_name = "music"
|
||||||
|
@ -83,13 +80,4 @@
|
||||||
name="retrieve_update_delete_author",
|
name="retrieve_update_delete_author",
|
||||||
),
|
),
|
||||||
path("anon/create/", CreateAnonMusicUserAPIView.as_view(), name="create-anon"),
|
path("anon/create/", CreateAnonMusicUserAPIView.as_view(), name="create-anon"),
|
||||||
path("search/", SearchAllAPIView.as_view(), name="search_all"),
|
|
||||||
path(
|
|
||||||
"drafts/callback/<uuid:token>/",
|
|
||||||
MusicDraftCallbackView.as_view(),
|
|
||||||
name="draft-callback",
|
|
||||||
),
|
|
||||||
path(
|
|
||||||
"drafts/<str:file_token>/", MusicDraftDetailView.as_view(), name="draft-detail"
|
|
||||||
),
|
|
||||||
]
|
]
|
||||||
|
|
|
@ -1,13 +1,11 @@
|
||||||
from drf_spectacular.utils import OpenApiExample, OpenApiParameter, extend_schema
|
from drf_spectacular.utils import OpenApiExample, OpenApiParameter, extend_schema
|
||||||
from rest_framework import generics, permissions, status
|
from rest_framework import generics, permissions
|
||||||
from rest_framework.generics import get_object_or_404
|
|
||||||
from rest_framework.response import Response
|
from rest_framework.response import Response
|
||||||
|
|
||||||
from akarpov.common.api.pagination import StandardResultsSetPagination
|
from akarpov.common.api.pagination import StandardResultsSetPagination
|
||||||
from akarpov.common.api.permissions import IsAdminOrReadOnly, IsCreatorOrReadOnly
|
from akarpov.common.api.permissions import IsAdminOrReadOnly, IsCreatorOrReadOnly
|
||||||
from akarpov.music.api.serializers import (
|
from akarpov.music.api.serializers import (
|
||||||
AddSongToPlaylistSerializer,
|
AddSongToPlaylistSerializer,
|
||||||
AllSearchSerializer,
|
|
||||||
AnonMusicUserSerializer,
|
AnonMusicUserSerializer,
|
||||||
FullAlbumSerializer,
|
FullAlbumSerializer,
|
||||||
FullAuthorSerializer,
|
FullAuthorSerializer,
|
||||||
|
@ -19,23 +17,19 @@
|
||||||
ListPlaylistSerializer,
|
ListPlaylistSerializer,
|
||||||
ListSongSerializer,
|
ListSongSerializer,
|
||||||
ListSongSlugsSerializer,
|
ListSongSlugsSerializer,
|
||||||
MusicDraftCallbackSerializer,
|
|
||||||
MusicDraftSerializer,
|
|
||||||
PlaylistSerializer,
|
PlaylistSerializer,
|
||||||
SongSerializer,
|
SongSerializer,
|
||||||
)
|
)
|
||||||
from akarpov.music.models import (
|
from akarpov.music.models import (
|
||||||
Album,
|
Album,
|
||||||
Author,
|
Author,
|
||||||
MusicDraft,
|
|
||||||
Playlist,
|
Playlist,
|
||||||
Song,
|
Song,
|
||||||
SongUserRating,
|
SongUserRating,
|
||||||
UserListenHistory,
|
UserListenHistory,
|
||||||
)
|
)
|
||||||
from akarpov.music.services.search import search_album, search_author, search_song
|
from akarpov.music.services.search import search_song
|
||||||
from akarpov.music.tasks import listen_to_song, process_draft_callback
|
from akarpov.music.tasks import listen_to_song
|
||||||
from akarpov.users.models import User
|
|
||||||
|
|
||||||
|
|
||||||
class LikedSongsContextMixin(generics.GenericAPIView):
|
class LikedSongsContextMixin(generics.GenericAPIView):
|
||||||
|
@ -357,25 +351,7 @@ class ListAlbumsAPIView(generics.ListAPIView):
|
||||||
serializer_class = ListAlbumSerializer
|
serializer_class = ListAlbumSerializer
|
||||||
pagination_class = StandardResultsSetPagination
|
pagination_class = StandardResultsSetPagination
|
||||||
permission_classes = [permissions.AllowAny]
|
permission_classes = [permissions.AllowAny]
|
||||||
|
queryset = Album.objects.cache().all()
|
||||||
def get_queryset(self):
|
|
||||||
search = self.request.query_params.get("search", None)
|
|
||||||
if search:
|
|
||||||
return search_album(search)
|
|
||||||
return Album.objects.cache().all()
|
|
||||||
|
|
||||||
@extend_schema(
|
|
||||||
parameters=[
|
|
||||||
OpenApiParameter(
|
|
||||||
name="search",
|
|
||||||
description="Search query for albums",
|
|
||||||
required=False,
|
|
||||||
type=str,
|
|
||||||
),
|
|
||||||
]
|
|
||||||
)
|
|
||||||
def get(self, request, *args, **kwargs):
|
|
||||||
return super().get(request, *args, **kwargs)
|
|
||||||
|
|
||||||
|
|
||||||
class RetrieveUpdateDestroyAlbumAPIView(
|
class RetrieveUpdateDestroyAlbumAPIView(
|
||||||
|
@ -392,25 +368,7 @@ class ListAuthorsAPIView(generics.ListAPIView):
|
||||||
serializer_class = ListAuthorSerializer
|
serializer_class = ListAuthorSerializer
|
||||||
pagination_class = StandardResultsSetPagination
|
pagination_class = StandardResultsSetPagination
|
||||||
permission_classes = [permissions.AllowAny]
|
permission_classes = [permissions.AllowAny]
|
||||||
|
queryset = Author.objects.cache().all()
|
||||||
def get_queryset(self):
|
|
||||||
search = self.request.query_params.get("search", None)
|
|
||||||
if search:
|
|
||||||
return search_author(search)
|
|
||||||
return Author.objects.cache().all()
|
|
||||||
|
|
||||||
@extend_schema(
|
|
||||||
parameters=[
|
|
||||||
OpenApiParameter(
|
|
||||||
name="search",
|
|
||||||
description="Search query for authors",
|
|
||||||
required=False,
|
|
||||||
type=str,
|
|
||||||
),
|
|
||||||
]
|
|
||||||
)
|
|
||||||
def get(self, request, *args, **kwargs):
|
|
||||||
return super().get(request, *args, **kwargs)
|
|
||||||
|
|
||||||
|
|
||||||
class RetrieveUpdateDestroyAuthorAPIView(
|
class RetrieveUpdateDestroyAuthorAPIView(
|
||||||
|
@ -433,27 +391,12 @@ def get_queryset(self):
|
||||||
def post(self, request, *args, **kwargs):
|
def post(self, request, *args, **kwargs):
|
||||||
serializer = self.get_serializer(data=request.data)
|
serializer = self.get_serializer(data=request.data)
|
||||||
serializer.is_valid(raise_exception=False)
|
serializer.is_valid(raise_exception=False)
|
||||||
|
data = serializer.validated_data
|
||||||
|
|
||||||
try:
|
try:
|
||||||
song = Song.objects.cache().get(slug=self.request.data.get("song", ""))
|
song = Song.objects.cache().get(slug=data["song"])
|
||||||
except Song.DoesNotExist:
|
except Song.DoesNotExist:
|
||||||
return Response(status=404)
|
return Response(status=404)
|
||||||
|
|
||||||
try:
|
|
||||||
user_id = self.request.data.get("user_id", None)
|
|
||||||
if user_id:
|
|
||||||
user_id_int = None
|
|
||||||
try:
|
|
||||||
user_id_int = int(user_id)
|
|
||||||
except ValueError:
|
|
||||||
...
|
|
||||||
if user_id_int:
|
|
||||||
user = User.objects.cache().get(id=user_id_int)
|
|
||||||
if user != self.request.user:
|
|
||||||
return Response(status=403)
|
|
||||||
except User.DoesNotExist:
|
|
||||||
...
|
|
||||||
|
|
||||||
if self.request.user.is_authenticated:
|
if self.request.user.is_authenticated:
|
||||||
listen_to_song.apply_async(
|
listen_to_song.apply_async(
|
||||||
kwargs={
|
kwargs={
|
||||||
|
@ -463,11 +406,11 @@ def post(self, request, *args, **kwargs):
|
||||||
},
|
},
|
||||||
countdown=2,
|
countdown=2,
|
||||||
)
|
)
|
||||||
elif "user_id" in self.request.data:
|
elif "user_id" in data:
|
||||||
listen_to_song.apply_async(
|
listen_to_song.apply_async(
|
||||||
kwargs={
|
kwargs={
|
||||||
"song_id": song.id,
|
"song_id": song.id,
|
||||||
"user_id": self.request.data.get("user_id", None),
|
"user_id": data["user_id"],
|
||||||
"anon": True,
|
"anon": True,
|
||||||
},
|
},
|
||||||
countdown=2,
|
countdown=2,
|
||||||
|
@ -496,88 +439,3 @@ def get_queryset(self):
|
||||||
class CreateAnonMusicUserAPIView(generics.CreateAPIView):
|
class CreateAnonMusicUserAPIView(generics.CreateAPIView):
|
||||||
serializer_class = AnonMusicUserSerializer
|
serializer_class = AnonMusicUserSerializer
|
||||||
permission_classes = [permissions.AllowAny]
|
permission_classes = [permissions.AllowAny]
|
||||||
|
|
||||||
|
|
||||||
class SearchAllAPIView(LikedSongsContextMixin, generics.GenericAPIView):
|
|
||||||
permission_classes = [permissions.AllowAny]
|
|
||||||
serializer_class = AllSearchSerializer
|
|
||||||
|
|
||||||
def get_serializer_context(self):
|
|
||||||
context = super().get_serializer_context()
|
|
||||||
context["request"] = self.request
|
|
||||||
return context
|
|
||||||
|
|
||||||
@extend_schema(
|
|
||||||
parameters=[
|
|
||||||
OpenApiParameter(
|
|
||||||
name="query",
|
|
||||||
description="Search query",
|
|
||||||
required=True,
|
|
||||||
type=str,
|
|
||||||
),
|
|
||||||
],
|
|
||||||
responses={
|
|
||||||
200: AllSearchSerializer,
|
|
||||||
},
|
|
||||||
)
|
|
||||||
def get(self, request, *args, **kwargs):
|
|
||||||
query = request.query_params.get("query", "").strip()
|
|
||||||
if not query:
|
|
||||||
return Response({"songs": [], "albums": [], "authors": []})
|
|
||||||
|
|
||||||
songs = search_song(query)[:10] # Top 10 songs
|
|
||||||
albums = search_album(query)[:5] # Top 5 albums
|
|
||||||
authors = search_author(query)[:5] # Top 5 authors
|
|
||||||
|
|
||||||
song_serializer = ListSongSerializer(
|
|
||||||
songs, many=True, context=self.get_serializer_context()
|
|
||||||
)
|
|
||||||
album_serializer = ListAlbumSerializer(
|
|
||||||
albums, many=True, context=self.get_serializer_context()
|
|
||||||
)
|
|
||||||
author_serializer = ListAuthorSerializer(
|
|
||||||
authors, many=True, context=self.get_serializer_context()
|
|
||||||
)
|
|
||||||
|
|
||||||
return Response(
|
|
||||||
{
|
|
||||||
"songs": song_serializer.data,
|
|
||||||
"albums": album_serializer.data,
|
|
||||||
"authors": author_serializer.data,
|
|
||||||
}
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class MusicDraftCallbackView(generics.GenericAPIView):
|
|
||||||
serializer_class = MusicDraftCallbackSerializer
|
|
||||||
|
|
||||||
@extend_schema(
|
|
||||||
description="Callback endpoint for external music service",
|
|
||||||
parameters=[
|
|
||||||
OpenApiParameter(
|
|
||||||
name="token",
|
|
||||||
type=str,
|
|
||||||
location=OpenApiParameter.PATH,
|
|
||||||
description="Draft callback token",
|
|
||||||
),
|
|
||||||
],
|
|
||||||
)
|
|
||||||
def post(self, request, token):
|
|
||||||
draft = get_object_or_404(MusicDraft, callback_token=token)
|
|
||||||
serializer = self.get_serializer(data=request.data)
|
|
||||||
serializer.is_valid(raise_exception=True)
|
|
||||||
|
|
||||||
process_draft_callback.delay(
|
|
||||||
draft_id=str(draft.id),
|
|
||||||
status=serializer.validated_data["status"],
|
|
||||||
meta_data=serializer.validated_data.get("meta_data"),
|
|
||||||
error_message=serializer.validated_data.get("error_message"),
|
|
||||||
)
|
|
||||||
|
|
||||||
return Response(status=status.HTTP_202_ACCEPTED)
|
|
||||||
|
|
||||||
|
|
||||||
class MusicDraftDetailView(generics.RetrieveAPIView):
|
|
||||||
queryset = MusicDraft.objects.all()
|
|
||||||
serializer_class = MusicDraftSerializer
|
|
||||||
lookup_field = "file_token"
|
|
||||||
|
|
|
@ -1,7 +1,7 @@
|
||||||
from django_elasticsearch_dsl import Document, fields
|
from django_elasticsearch_dsl import Document, fields
|
||||||
from django_elasticsearch_dsl.registries import registry
|
from django_elasticsearch_dsl.registries import registry
|
||||||
|
|
||||||
from akarpov.music.models import Album, Author, Song
|
from akarpov.music.models import Song
|
||||||
|
|
||||||
|
|
||||||
@registry.register_document
|
@registry.register_document
|
||||||
|
@ -14,12 +14,6 @@ class SongDocument(Document):
|
||||||
"raw": fields.KeywordField(normalizer="lowercase"),
|
"raw": fields.KeywordField(normalizer="lowercase"),
|
||||||
},
|
},
|
||||||
),
|
),
|
||||||
"name_transliterated": fields.TextField(
|
|
||||||
analyzer="transliterate",
|
|
||||||
fields={
|
|
||||||
"raw": fields.KeywordField(),
|
|
||||||
},
|
|
||||||
),
|
|
||||||
"link": fields.TextField(),
|
"link": fields.TextField(),
|
||||||
"meta": fields.ObjectField(dynamic=True),
|
"meta": fields.ObjectField(dynamic=True),
|
||||||
},
|
},
|
||||||
|
@ -33,12 +27,6 @@ class SongDocument(Document):
|
||||||
"raw": fields.KeywordField(normalizer="lowercase"),
|
"raw": fields.KeywordField(normalizer="lowercase"),
|
||||||
},
|
},
|
||||||
),
|
),
|
||||||
"name_transliterated": fields.TextField(
|
|
||||||
analyzer="transliterate",
|
|
||||||
fields={
|
|
||||||
"raw": fields.KeywordField(),
|
|
||||||
},
|
|
||||||
),
|
|
||||||
"link": fields.TextField(),
|
"link": fields.TextField(),
|
||||||
"meta": fields.ObjectField(dynamic=True),
|
"meta": fields.ObjectField(dynamic=True),
|
||||||
},
|
},
|
||||||
|
@ -47,18 +35,9 @@ class SongDocument(Document):
|
||||||
name = fields.TextField(
|
name = fields.TextField(
|
||||||
attr="name",
|
attr="name",
|
||||||
fields={
|
fields={
|
||||||
"raw": fields.KeywordField(),
|
"raw": fields.KeywordField(normalizer="lowercase"),
|
||||||
"exact": fields.KeywordField(normalizer="lowercase"),
|
|
||||||
},
|
},
|
||||||
)
|
)
|
||||||
name_transliterated = fields.TextField(
|
|
||||||
attr="name",
|
|
||||||
analyzer="transliterate",
|
|
||||||
fields={
|
|
||||||
"raw": fields.KeywordField(),
|
|
||||||
},
|
|
||||||
)
|
|
||||||
suggest = fields.CompletionField()
|
|
||||||
|
|
||||||
meta = fields.ObjectField(dynamic=True)
|
meta = fields.ObjectField(dynamic=True)
|
||||||
|
|
||||||
|
@ -69,17 +48,13 @@ class Index:
|
||||||
"number_of_replicas": 0,
|
"number_of_replicas": 0,
|
||||||
"analysis": {
|
"analysis": {
|
||||||
"filter": {
|
"filter": {
|
||||||
"my_transliterator": {
|
|
||||||
"type": "icu_transform",
|
|
||||||
"id": "Any-Latin; NFD; [:Nonspacing Mark:] Remove; NFC",
|
|
||||||
},
|
|
||||||
"russian_stop": {
|
"russian_stop": {
|
||||||
"type": "stop",
|
"type": "stop",
|
||||||
"stopwords": "_russian_",
|
"stopwords": "_russian_",
|
||||||
},
|
},
|
||||||
"russian_keywords": {
|
"russian_keywords": {
|
||||||
"type": "keyword_marker",
|
"type": "keyword_marker",
|
||||||
"keywords": ["песня", "музыка", "певец", "альбом"],
|
"keywords": ["пример"],
|
||||||
},
|
},
|
||||||
"russian_stemmer": {
|
"russian_stemmer": {
|
||||||
"type": "stemmer",
|
"type": "stemmer",
|
||||||
|
@ -105,13 +80,6 @@ class Index:
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
"analyzer": {
|
"analyzer": {
|
||||||
"transliterate": {
|
|
||||||
"tokenizer": "standard",
|
|
||||||
"filter": [
|
|
||||||
"lowercase",
|
|
||||||
"my_transliterator",
|
|
||||||
],
|
|
||||||
},
|
|
||||||
"russian": {
|
"russian": {
|
||||||
"tokenizer": "standard",
|
"tokenizer": "standard",
|
||||||
"filter": [
|
"filter": [
|
||||||
|
@ -169,74 +137,3 @@ def get_instances_from_related(self, related_instance):
|
||||||
if isinstance(related_instance, Song):
|
if isinstance(related_instance, Song):
|
||||||
return related_instance.album
|
return related_instance.album
|
||||||
return related_instance.songs.all()
|
return related_instance.songs.all()
|
||||||
|
|
||||||
|
|
||||||
@registry.register_document
|
|
||||||
class AuthorDocument(Document):
|
|
||||||
name = fields.TextField(
|
|
||||||
fields={
|
|
||||||
"raw": fields.KeywordField(),
|
|
||||||
"exact": fields.KeywordField(normalizer="lowercase"),
|
|
||||||
},
|
|
||||||
)
|
|
||||||
name_transliterated = fields.TextField(
|
|
||||||
attr="name",
|
|
||||||
analyzer="transliterate",
|
|
||||||
fields={
|
|
||||||
"raw": fields.KeywordField(),
|
|
||||||
},
|
|
||||||
)
|
|
||||||
suggest = fields.CompletionField()
|
|
||||||
meta = fields.ObjectField(dynamic=True)
|
|
||||||
|
|
||||||
class Index:
|
|
||||||
name = "authors"
|
|
||||||
settings = SongDocument.Index.settings # Reuse settings
|
|
||||||
|
|
||||||
class Django:
|
|
||||||
model = Author
|
|
||||||
|
|
||||||
|
|
||||||
@registry.register_document
|
|
||||||
class AlbumDocument(Document):
|
|
||||||
name = fields.TextField(
|
|
||||||
fields={
|
|
||||||
"raw": fields.KeywordField(),
|
|
||||||
"exact": fields.KeywordField(normalizer="lowercase"),
|
|
||||||
},
|
|
||||||
)
|
|
||||||
name_transliterated = fields.TextField(
|
|
||||||
attr="name",
|
|
||||||
analyzer="transliterate",
|
|
||||||
fields={
|
|
||||||
"raw": fields.KeywordField(),
|
|
||||||
},
|
|
||||||
)
|
|
||||||
suggest = fields.CompletionField()
|
|
||||||
meta = fields.ObjectField(dynamic=True)
|
|
||||||
authors = fields.NestedField(
|
|
||||||
attr="authors",
|
|
||||||
properties={
|
|
||||||
"name": fields.TextField(
|
|
||||||
fields={
|
|
||||||
"raw": fields.KeywordField(normalizer="lowercase"),
|
|
||||||
},
|
|
||||||
),
|
|
||||||
"name_transliterated": fields.TextField(
|
|
||||||
attr="name",
|
|
||||||
analyzer="transliterate",
|
|
||||||
fields={
|
|
||||||
"raw": fields.KeywordField(),
|
|
||||||
},
|
|
||||||
),
|
|
||||||
"link": fields.TextField(),
|
|
||||||
"meta": fields.ObjectField(dynamic=True),
|
|
||||||
},
|
|
||||||
)
|
|
||||||
|
|
||||||
class Index:
|
|
||||||
name = "albums"
|
|
||||||
settings = SongDocument.Index.settings # Reuse settings
|
|
||||||
|
|
||||||
class Django:
|
|
||||||
model = Album
|
|
||||||
|
|
|
@ -1,89 +0,0 @@
|
||||||
# Generated by Django 4.2.16 on 2024-10-26 10:37
|
|
||||||
|
|
||||||
from django.db import migrations, models
|
|
||||||
import django.db.models.deletion
|
|
||||||
import uuid
|
|
||||||
|
|
||||||
|
|
||||||
class Migration(migrations.Migration):
|
|
||||||
dependencies = [
|
|
||||||
("music", "0016_anonmusicuser_song_created_song_volume_and_more"),
|
|
||||||
]
|
|
||||||
|
|
||||||
operations = [
|
|
||||||
migrations.CreateModel(
|
|
||||||
name="MusicDraft",
|
|
||||||
fields=[
|
|
||||||
(
|
|
||||||
"id",
|
|
||||||
models.UUIDField(
|
|
||||||
default=uuid.uuid4,
|
|
||||||
editable=False,
|
|
||||||
primary_key=True,
|
|
||||||
serialize=False,
|
|
||||||
),
|
|
||||||
),
|
|
||||||
(
|
|
||||||
"status",
|
|
||||||
models.CharField(
|
|
||||||
choices=[
|
|
||||||
("pending", "Pending"),
|
|
||||||
("processing", "Processing"),
|
|
||||||
("failed", "Failed"),
|
|
||||||
("complete", "Complete"),
|
|
||||||
],
|
|
||||||
default="pending",
|
|
||||||
max_length=20,
|
|
||||||
),
|
|
||||||
),
|
|
||||||
(
|
|
||||||
"provider",
|
|
||||||
models.CharField(
|
|
||||||
choices=[
|
|
||||||
("spotify", "Spotify"),
|
|
||||||
("yandex", "Yandex"),
|
|
||||||
("youtube", "YouTube"),
|
|
||||||
],
|
|
||||||
max_length=20,
|
|
||||||
),
|
|
||||||
),
|
|
||||||
("original_url", models.URLField()),
|
|
||||||
("meta_data", models.JSONField(blank=True, null=True)),
|
|
||||||
("file_token", models.CharField(max_length=100, unique=True)),
|
|
||||||
("created", models.DateTimeField(auto_now_add=True)),
|
|
||||||
("updated", models.DateTimeField(auto_now=True)),
|
|
||||||
("error_message", models.TextField(blank=True, null=True)),
|
|
||||||
("user_id", models.IntegerField(null=True)),
|
|
||||||
(
|
|
||||||
"callback_token",
|
|
||||||
models.UUIDField(default=uuid.uuid4, editable=False),
|
|
||||||
),
|
|
||||||
],
|
|
||||||
),
|
|
||||||
migrations.CreateModel(
|
|
||||||
name="MusicDraftFile",
|
|
||||||
fields=[
|
|
||||||
(
|
|
||||||
"id",
|
|
||||||
models.BigAutoField(
|
|
||||||
auto_created=True,
|
|
||||||
primary_key=True,
|
|
||||||
serialize=False,
|
|
||||||
verbose_name="ID",
|
|
||||||
),
|
|
||||||
),
|
|
||||||
("file", models.FileField(upload_to="music_drafts/")),
|
|
||||||
("original_name", models.CharField(max_length=255)),
|
|
||||||
("mime_type", models.CharField(max_length=100)),
|
|
||||||
("created", models.DateTimeField(auto_now_add=True)),
|
|
||||||
(
|
|
||||||
"draft",
|
|
||||||
models.ForeignKey(
|
|
||||||
on_delete=django.db.models.deletion.CASCADE,
|
|
||||||
related_name="files",
|
|
||||||
to="music.musicdraft",
|
|
||||||
),
|
|
||||||
),
|
|
||||||
],
|
|
||||||
),
|
|
||||||
]
|
|
|
@ -1,7 +1,6 @@
|
||||||
import uuid
|
import uuid
|
||||||
|
|
||||||
from django.contrib.postgres.fields import ArrayField
|
from django.contrib.postgres.fields import ArrayField
|
||||||
from django.contrib.sites.models import Site
|
|
||||||
from django.db import models
|
from django.db import models
|
||||||
from django.urls import reverse
|
from django.urls import reverse
|
||||||
|
|
||||||
|
@ -87,11 +86,6 @@ def album_name(self):
|
||||||
def artists_names(self):
|
def artists_names(self):
|
||||||
return cache_model_property(self, "_authors_names")
|
return cache_model_property(self, "_authors_names")
|
||||||
|
|
||||||
def get_first_author_name(self):
|
|
||||||
if self.authors:
|
|
||||||
return self.authors.first().name
|
|
||||||
return ""
|
|
||||||
|
|
||||||
def __str__(self):
|
def __str__(self):
|
||||||
return self.name
|
return self.name
|
||||||
|
|
||||||
|
@ -99,50 +93,6 @@ class SlugMeta:
|
||||||
slug_length = 10
|
slug_length = 10
|
||||||
|
|
||||||
|
|
||||||
class MusicDraft(models.Model):
|
|
||||||
STATUS_CHOICES = (
|
|
||||||
("pending", "Pending"),
|
|
||||||
("processing", "Processing"),
|
|
||||||
("failed", "Failed"),
|
|
||||||
("complete", "Complete"),
|
|
||||||
)
|
|
||||||
|
|
||||||
PROVIDER_CHOICES = (
|
|
||||||
("spotify", "Spotify"),
|
|
||||||
("yandex", "Yandex"),
|
|
||||||
("youtube", "YouTube"),
|
|
||||||
)
|
|
||||||
|
|
||||||
id = models.UUIDField(primary_key=True, default=uuid.uuid4, editable=False)
|
|
||||||
status = models.CharField(max_length=20, choices=STATUS_CHOICES, default="pending")
|
|
||||||
provider = models.CharField(max_length=20, choices=PROVIDER_CHOICES)
|
|
||||||
original_url = models.URLField()
|
|
||||||
meta_data = models.JSONField(null=True, blank=True)
|
|
||||||
file_token = models.CharField(max_length=100, unique=True)
|
|
||||||
created = models.DateTimeField(auto_now_add=True)
|
|
||||||
updated = models.DateTimeField(auto_now=True)
|
|
||||||
error_message = models.TextField(null=True, blank=True)
|
|
||||||
user_id = models.IntegerField(null=True)
|
|
||||||
callback_token = models.UUIDField(default=uuid.uuid4, editable=False)
|
|
||||||
|
|
||||||
def get_callback_url(self):
|
|
||||||
site = Site.objects.get_current()
|
|
||||||
path = reverse(
|
|
||||||
"music:api:draft-callback", kwargs={"token": self.callback_token}
|
|
||||||
)
|
|
||||||
return f"https://{site.domain}{path}"
|
|
||||||
|
|
||||||
|
|
||||||
class MusicDraftFile(models.Model):
|
|
||||||
draft = models.ForeignKey(
|
|
||||||
MusicDraft, on_delete=models.CASCADE, related_name="files"
|
|
||||||
)
|
|
||||||
file = models.FileField(upload_to="music_drafts/")
|
|
||||||
original_name = models.CharField(max_length=255)
|
|
||||||
mime_type = models.CharField(max_length=100)
|
|
||||||
created = models.DateTimeField(auto_now_add=True)
|
|
||||||
|
|
||||||
|
|
||||||
class Playlist(ShortLinkModel, UserHistoryModel):
|
class Playlist(ShortLinkModel, UserHistoryModel):
|
||||||
name = models.CharField(max_length=200)
|
name = models.CharField(max_length=200)
|
||||||
private = models.BooleanField(default=True)
|
private = models.BooleanField(default=True)
|
||||||
|
|
|
@ -2,11 +2,7 @@
|
||||||
import re
|
import re
|
||||||
|
|
||||||
import requests
|
import requests
|
||||||
|
from deep_translator import GoogleTranslator
|
||||||
try:
|
|
||||||
from deep_translator import GoogleTranslator # TODO: move to another service
|
|
||||||
except requests.exceptions.JSONDecodeError:
|
|
||||||
print("Failed to initialize GoogleTranslator due to external API issues.")
|
|
||||||
from django.core.files import File
|
from django.core.files import File
|
||||||
from django.db import transaction
|
from django.db import transaction
|
||||||
from django.utils.text import slugify
|
from django.utils.text import slugify
|
||||||
|
@ -126,15 +122,12 @@ def load_track(
|
||||||
album=album,
|
album=album,
|
||||||
):
|
):
|
||||||
return sng.first()
|
return sng.first()
|
||||||
try:
|
|
||||||
if not path.endswith(".mp3"):
|
if not path.endswith(".mp3"):
|
||||||
mp3_path = path.replace(path.split(".")[-1], "mp3")
|
mp3_path = path.replace(path.split(".")[-1], "mp3")
|
||||||
AudioSegment.from_file(path).export(mp3_path)
|
AudioSegment.from_file(path).export(mp3_path)
|
||||||
os.remove(path)
|
os.remove(path)
|
||||||
path = mp3_path
|
path = mp3_path
|
||||||
except Exception as e:
|
|
||||||
print(e)
|
|
||||||
return Song.objects.none()
|
|
||||||
|
|
||||||
tag = MP3(path, ID3=ID3)
|
tag = MP3(path, ID3=ID3)
|
||||||
|
|
||||||
|
|
|
@ -1,84 +0,0 @@
|
||||||
import os
|
|
||||||
|
|
||||||
from django.core.files import File
|
|
||||||
|
|
||||||
from akarpov.music.models import Album, Author, MusicDraft, Song
|
|
||||||
|
|
||||||
|
|
||||||
def save_song_from_draft(draft: MusicDraft) -> Song | None:
|
|
||||||
"""
|
|
||||||
Create a Song instance from a completed MusicDraft
|
|
||||||
"""
|
|
||||||
try:
|
|
||||||
if not draft.files.exists():
|
|
||||||
draft.status = "failed"
|
|
||||||
draft.error_message = "No files associated with draft"
|
|
||||||
draft.save()
|
|
||||||
return None
|
|
||||||
|
|
||||||
# Get the music file
|
|
||||||
draft_file = draft.files.filter(mime_type__startswith="audio/").first()
|
|
||||||
if not draft_file:
|
|
||||||
draft.status = "failed"
|
|
||||||
draft.error_message = "No audio file found in draft"
|
|
||||||
draft.save()
|
|
||||||
return None
|
|
||||||
|
|
||||||
# Get metadata from draft
|
|
||||||
meta_data = draft.meta_data or {}
|
|
||||||
|
|
||||||
# Create song instance
|
|
||||||
song = Song(
|
|
||||||
name=meta_data.get("title", ""),
|
|
||||||
length=meta_data.get("length"),
|
|
||||||
link=draft.original_url,
|
|
||||||
meta=meta_data,
|
|
||||||
)
|
|
||||||
|
|
||||||
# Handle album
|
|
||||||
if "album" in meta_data:
|
|
||||||
album, _ = Album.objects.get_or_create(
|
|
||||||
name=meta_data["album"].get("name", ""),
|
|
||||||
defaults={"meta": meta_data["album"]},
|
|
||||||
)
|
|
||||||
song.album = album
|
|
||||||
|
|
||||||
# Save song to get an ID
|
|
||||||
with open(draft_file.file.path, "rb") as f:
|
|
||||||
song.file.save(
|
|
||||||
os.path.basename(draft_file.original_name), File(f), save=True
|
|
||||||
)
|
|
||||||
|
|
||||||
# Handle authors
|
|
||||||
if "artists" in meta_data:
|
|
||||||
authors = []
|
|
||||||
for artist_data in meta_data["artists"]:
|
|
||||||
author, _ = Author.objects.get_or_create(
|
|
||||||
name=artist_data.get("name", ""), defaults={"meta": artist_data}
|
|
||||||
)
|
|
||||||
authors.append(author)
|
|
||||||
song.authors.set(authors)
|
|
||||||
|
|
||||||
# Handle image if present
|
|
||||||
image_file = draft.files.filter(mime_type__startswith="image/").first()
|
|
||||||
if image_file:
|
|
||||||
with open(image_file.file.path, "rb") as f:
|
|
||||||
song.image.save(
|
|
||||||
os.path.basename(image_file.original_name), File(f), save=True
|
|
||||||
)
|
|
||||||
|
|
||||||
# Add user if specified
|
|
||||||
if draft.user_id:
|
|
||||||
song.creator_id = draft.user_id
|
|
||||||
song.save()
|
|
||||||
|
|
||||||
# Clean up draft files
|
|
||||||
draft.delete()
|
|
||||||
|
|
||||||
return song
|
|
||||||
|
|
||||||
except Exception as e:
|
|
||||||
draft.status = "failed"
|
|
||||||
draft.error_message = str(e)
|
|
||||||
draft.save()
|
|
||||||
return None
|
|
|
@ -3,15 +3,10 @@
|
||||||
|
|
||||||
import requests
|
import requests
|
||||||
import spotipy
|
import spotipy
|
||||||
|
from deep_translator import GoogleTranslator
|
||||||
try:
|
|
||||||
from deep_translator import GoogleTranslator
|
|
||||||
except requests.exceptions.JSONDecodeError:
|
|
||||||
print("Failed to initialize GoogleTranslator due to external API issues.")
|
|
||||||
from django.conf import settings
|
from django.conf import settings
|
||||||
from django.core.files import File
|
from django.core.files import File
|
||||||
from django.db import transaction
|
from django.db import transaction
|
||||||
from django.db.models import Model
|
|
||||||
from django.utils.text import slugify
|
from django.utils.text import slugify
|
||||||
from spotipy import SpotifyClientCredentials
|
from spotipy import SpotifyClientCredentials
|
||||||
from yandex_music import Client, Cover
|
from yandex_music import Client, Cover
|
||||||
|
@ -22,33 +17,30 @@
|
||||||
from akarpov.utils.text import is_similar_artist, normalize_text
|
from akarpov.utils.text import is_similar_artist, normalize_text
|
||||||
|
|
||||||
|
|
||||||
def generate_readable_slug(name: str, model: Model) -> str:
|
def generate_readable_slug(name: str, model) -> str:
|
||||||
# Translate and slugify the name
|
# Translate and slugify the name
|
||||||
slug = safe_translate(name)
|
slug = str(
|
||||||
|
slugify(
|
||||||
|
GoogleTranslator(source="auto", target="en").translate(
|
||||||
|
name,
|
||||||
|
target_language="en",
|
||||||
|
)
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|
||||||
# Truncate slug if it's too long
|
|
||||||
if len(slug) > 20:
|
if len(slug) > 20:
|
||||||
slug = slug[:20]
|
slug = slug[:20]
|
||||||
last_dash = slug.rfind("-")
|
last_dash = slug.rfind("-")
|
||||||
if last_dash != -1:
|
if last_dash != -1:
|
||||||
slug = slug[:last_dash]
|
slug = slug[:last_dash]
|
||||||
|
|
||||||
original_slug = slug
|
|
||||||
|
|
||||||
# Ensure uniqueness
|
|
||||||
counter = 1
|
|
||||||
while model.objects.filter(slug=slug).exists():
|
while model.objects.filter(slug=slug).exists():
|
||||||
if len(original_slug) > 14:
|
if len(slug) > 14:
|
||||||
truncated_slug = original_slug[:14]
|
slug = slug[:14]
|
||||||
last_dash = truncated_slug.rfind("-")
|
last_dash = slug.rfind("-")
|
||||||
if last_dash != -1:
|
if last_dash != -1:
|
||||||
truncated_slug = truncated_slug[:last_dash]
|
slug = slug[:last_dash]
|
||||||
else:
|
slug = slug + "_" + generate_charset(5)
|
||||||
truncated_slug = original_slug
|
|
||||||
|
|
||||||
suffix = f"_{generate_charset(5)}" if counter == 1 else f"_{counter}"
|
|
||||||
slug = f"{truncated_slug}{suffix}"
|
|
||||||
counter += 1
|
|
||||||
|
|
||||||
return slug
|
return slug
|
||||||
|
|
||||||
|
@ -221,12 +213,16 @@ def update_album_info(album: AlbumModel, author_name: str = None) -> None:
|
||||||
client = yandex_login()
|
client = yandex_login()
|
||||||
spotify_session = create_spotify_session()
|
spotify_session = create_spotify_session()
|
||||||
|
|
||||||
search_term = f"{album.name} - {author_name}" if author_name else album.name
|
if author_name:
|
||||||
|
yandex_album_info = get_yandex_album_info(
|
||||||
yandex_album_info = get_api_info(get_yandex_album_info, search_term, client)
|
album.name + " - " + author_name, client
|
||||||
spotify_album_info = get_api_info(
|
)
|
||||||
get_spotify_album_info, search_term, spotify_session
|
spotify_album_info = get_spotify_album_info(
|
||||||
)
|
album.name + " - " + author_name, spotify_session
|
||||||
|
)
|
||||||
|
else:
|
||||||
|
yandex_album_info = get_yandex_album_info(album.name, client)
|
||||||
|
spotify_album_info = get_spotify_album_info(album.name, spotify_session)
|
||||||
|
|
||||||
# Combine and prioritize Spotify data
|
# Combine and prioritize Spotify data
|
||||||
album_data = {}
|
album_data = {}
|
||||||
|
@ -236,14 +232,14 @@ def update_album_info(album: AlbumModel, author_name: str = None) -> None:
|
||||||
"name": spotify_album_info.get("name", album.name),
|
"name": spotify_album_info.get("name", album.name),
|
||||||
"release_date": spotify_album_info.get("release_date", ""),
|
"release_date": spotify_album_info.get("release_date", ""),
|
||||||
"total_tracks": spotify_album_info.get("total_tracks", ""),
|
"total_tracks": spotify_album_info.get("total_tracks", ""),
|
||||||
"link": spotify_album_info.get("external_urls", {}).get("spotify", ""),
|
"link": spotify_album_info["external_urls"]["spotify"],
|
||||||
"genre": spotify_album_info.get("genres", []),
|
"genre": spotify_album_info.get("genres", []),
|
||||||
}
|
}
|
||||||
if yandex_album_info:
|
if yandex_album_info:
|
||||||
album_data.update(
|
album_data.update(
|
||||||
{
|
{
|
||||||
"name": album_data.get("name") or yandex_album_info.title,
|
"name": album_data.get("name", yandex_album_info.title),
|
||||||
"genre": album_data.get("genre") or yandex_album_info.genre,
|
"genre": album_data.get("genre", yandex_album_info.genre),
|
||||||
"description": yandex_album_info.description,
|
"description": yandex_album_info.description,
|
||||||
"type": yandex_album_info.type,
|
"type": yandex_album_info.type,
|
||||||
}
|
}
|
||||||
|
@ -253,120 +249,102 @@ def update_album_info(album: AlbumModel, author_name: str = None) -> None:
|
||||||
album.save()
|
album.save()
|
||||||
|
|
||||||
# Handle Album Image - Prefer Spotify, fallback to Yandex
|
# Handle Album Image - Prefer Spotify, fallback to Yandex
|
||||||
image_path = get_album_image(spotify_album_info, yandex_album_info)
|
image_path = None
|
||||||
|
if (
|
||||||
|
spotify_album_info
|
||||||
|
and "images" in spotify_album_info
|
||||||
|
and spotify_album_info["images"]
|
||||||
|
):
|
||||||
|
image_path = download_image(
|
||||||
|
spotify_album_info["images"][0]["url"], settings.MEDIA_ROOT
|
||||||
|
)
|
||||||
|
elif yandex_album_info and yandex_album_info.cover_uri:
|
||||||
|
image_path = download_image(
|
||||||
|
"https://" + yandex_album_info.cover_uri, settings.MEDIA_ROOT
|
||||||
|
)
|
||||||
|
|
||||||
|
generated_name = slugify(
|
||||||
|
GoogleTranslator(source="auto", target="en").translate(
|
||||||
|
album.name,
|
||||||
|
target_language="en",
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|
||||||
if image_path:
|
if image_path:
|
||||||
save_album_image(album, image_path)
|
|
||||||
|
|
||||||
# Update Album Authors from Spotify data if available
|
|
||||||
if spotify_album_info and "artists" in spotify_album_info:
|
|
||||||
update_album_authors(album, spotify_album_info["artists"])
|
|
||||||
|
|
||||||
album.slug = generate_readable_slug(album.name, AlbumModel)
|
|
||||||
album.save()
|
|
||||||
|
|
||||||
|
|
||||||
def get_album_image(spotify_info, yandex_info):
|
|
||||||
if spotify_info and "images" in spotify_info and spotify_info["images"]:
|
|
||||||
return download_image(spotify_info["images"][0]["url"], settings.MEDIA_ROOT)
|
|
||||||
elif yandex_info and yandex_info.cover_uri:
|
|
||||||
return download_image("https://" + yandex_info.cover_uri, settings.MEDIA_ROOT)
|
|
||||||
return None
|
|
||||||
|
|
||||||
|
|
||||||
def save_album_image(album, image_path):
|
|
||||||
if not image_path:
|
|
||||||
return
|
|
||||||
|
|
||||||
try:
|
|
||||||
generated_name = safe_translate(album.name)
|
|
||||||
with open(image_path, "rb") as f:
|
with open(image_path, "rb") as f:
|
||||||
album.image.save(
|
album.image.save(
|
||||||
generated_name + ".png",
|
generated_name + ".png",
|
||||||
File(f, name=generated_name + ".png"),
|
File(
|
||||||
|
f,
|
||||||
|
name=generated_name + ".png",
|
||||||
|
),
|
||||||
save=True,
|
save=True,
|
||||||
)
|
)
|
||||||
os.remove(image_path)
|
os.remove(image_path)
|
||||||
album.save()
|
album.save()
|
||||||
except Exception as e:
|
|
||||||
print(f"Error saving album image: {str(e)}")
|
|
||||||
|
|
||||||
|
# Update Album Authors from Spotify data if available
|
||||||
|
if spotify_album_info and "artists" in spotify_album_info:
|
||||||
|
album_authors = []
|
||||||
|
for artist in spotify_album_info["artists"]:
|
||||||
|
author, created = Author.objects.get_or_create(name=artist["name"])
|
||||||
|
album_authors.append(author)
|
||||||
|
album.authors.set(album_authors)
|
||||||
|
|
||||||
def update_album_authors(album, artists):
|
album.slug = generate_readable_slug(album.name, AlbumModel)
|
||||||
album_authors = []
|
album.save()
|
||||||
for artist in artists:
|
|
||||||
author, created = Author.objects.get_or_create(name=artist["name"])
|
|
||||||
album_authors.append(author)
|
|
||||||
album.authors.set(album_authors)
|
|
||||||
|
|
||||||
|
|
||||||
def update_author_info(author: Author) -> None:
|
def update_author_info(author: Author) -> None:
|
||||||
client = yandex_login()
|
client = yandex_login()
|
||||||
spotify_session = create_spotify_session()
|
spotify_session = create_spotify_session()
|
||||||
|
|
||||||
yandex_artist_info = get_api_info(get_yandex_artist_info, author.name, client)
|
# Retrieve info from both services
|
||||||
spotify_artist_info = get_api_info(
|
yandex_artist_info = get_yandex_artist_info(author.name, client)
|
||||||
get_spotify_artist_info, author.name, spotify_session
|
spotify_artist_info = get_spotify_artist_info(author.name, spotify_session)
|
||||||
)
|
|
||||||
|
|
||||||
author_data = combine_artist_data(author, spotify_artist_info, yandex_artist_info)
|
# Combine and prioritize Spotify data
|
||||||
|
|
||||||
with transaction.atomic():
|
|
||||||
author.meta = author_data
|
|
||||||
author.save()
|
|
||||||
|
|
||||||
image_path = get_author_image(spotify_artist_info, yandex_artist_info)
|
|
||||||
|
|
||||||
if image_path:
|
|
||||||
save_author_image(author, image_path)
|
|
||||||
|
|
||||||
author.slug = generate_readable_slug(author.name, Author)
|
|
||||||
with transaction.atomic():
|
|
||||||
author.save()
|
|
||||||
|
|
||||||
|
|
||||||
def get_api_info(api_func, search_term, session):
|
|
||||||
try:
|
|
||||||
return api_func(search_term, session)
|
|
||||||
except Exception as e:
|
|
||||||
print(f"Error fetching info from {api_func.__name__}: {str(e)}")
|
|
||||||
return None
|
|
||||||
|
|
||||||
|
|
||||||
def combine_artist_data(author, spotify_info, yandex_info):
|
|
||||||
author_data = {}
|
author_data = {}
|
||||||
if spotify_info:
|
if spotify_artist_info:
|
||||||
author_data = {
|
author_data = {
|
||||||
"name": spotify_info.get("name", author.name),
|
"name": spotify_artist_info.get("name", author.name),
|
||||||
"genres": spotify_info.get("genres", []),
|
"genres": spotify_artist_info.get("genres", []),
|
||||||
"popularity": spotify_info.get("popularity", 0),
|
"popularity": spotify_artist_info.get("popularity", 0),
|
||||||
"link": spotify_info.get("external_urls", {}).get("spotify", ""),
|
"link": spotify_artist_info["external_urls"]["spotify"],
|
||||||
}
|
}
|
||||||
if yandex_info:
|
if yandex_artist_info:
|
||||||
author_data.update(
|
author_data.update(
|
||||||
{
|
{
|
||||||
"name": author_data.get("name") or yandex_info.name,
|
"name": author_data.get("name", yandex_artist_info.name),
|
||||||
"genres": author_data.get("genres") or yandex_info.genres,
|
"genres": author_data.get("genres", yandex_artist_info.genres),
|
||||||
"description": yandex_info.description,
|
"description": yandex_artist_info.description,
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
return author_data
|
|
||||||
|
|
||||||
|
author.meta = author_data
|
||||||
|
with transaction.atomic():
|
||||||
|
author.save()
|
||||||
|
|
||||||
def get_author_image(spotify_info, yandex_info):
|
# Handle Author Image - Prefer Spotify, fallback to Yandex
|
||||||
if spotify_info and "images" in spotify_info and spotify_info["images"]:
|
image_path = None
|
||||||
return download_image(spotify_info["images"][0]["url"], settings.MEDIA_ROOT)
|
if (
|
||||||
elif yandex_info and yandex_info.cover:
|
spotify_artist_info
|
||||||
return download_image(yandex_info.cover, settings.MEDIA_ROOT)
|
and "images" in spotify_artist_info
|
||||||
return None
|
and spotify_artist_info["images"]
|
||||||
|
):
|
||||||
|
image_path = download_image(
|
||||||
|
spotify_artist_info["images"][0]["url"], settings.MEDIA_ROOT
|
||||||
|
)
|
||||||
|
elif yandex_artist_info and yandex_artist_info.cover:
|
||||||
|
image_path = download_image(yandex_artist_info.cover, settings.MEDIA_ROOT)
|
||||||
|
|
||||||
|
generated_name = slugify(
|
||||||
def save_author_image(author, image_path):
|
GoogleTranslator(source="auto", target="en").translate(
|
||||||
if not image_path:
|
author.name,
|
||||||
return
|
target_language="en",
|
||||||
|
)
|
||||||
try:
|
)
|
||||||
generated_name = safe_translate(author.name)
|
if image_path:
|
||||||
with open(image_path, "rb") as f:
|
with open(image_path, "rb") as f:
|
||||||
author.image.save(
|
author.image.save(
|
||||||
generated_name + ".png",
|
generated_name + ".png",
|
||||||
|
@ -375,29 +353,21 @@ def save_author_image(author, image_path):
|
||||||
)
|
)
|
||||||
os.remove(image_path)
|
os.remove(image_path)
|
||||||
author.save()
|
author.save()
|
||||||
except Exception as e:
|
|
||||||
print(f"Error saving author image: {str(e)}")
|
|
||||||
|
|
||||||
|
author.slug = generate_readable_slug(author.name, Author)
|
||||||
def safe_translate(text):
|
with transaction.atomic():
|
||||||
try:
|
author.save()
|
||||||
translated = GoogleTranslator(source="auto", target="en").translate(text)
|
|
||||||
return slugify(translated)
|
|
||||||
except Exception as e:
|
|
||||||
print(f"Error translating text: {str(e)}")
|
|
||||||
return slugify(text)
|
|
||||||
|
|
||||||
|
|
||||||
def search_all_platforms(track_name: str) -> dict:
|
def search_all_platforms(track_name: str) -> dict:
|
||||||
print(track_name)
|
print(track_name)
|
||||||
# session = spotipy.Spotify(
|
session = spotipy.Spotify(
|
||||||
# auth_manager=spotipy.SpotifyClientCredentials(
|
auth_manager=spotipy.SpotifyClientCredentials(
|
||||||
# client_id=settings.MUSIC_SPOTIFY_ID,
|
client_id=settings.MUSIC_SPOTIFY_ID,
|
||||||
# client_secret=settings.MUSIC_SPOTIFY_SECRET,
|
client_secret=settings.MUSIC_SPOTIFY_SECRET,
|
||||||
# )
|
)
|
||||||
# )
|
)
|
||||||
# spotify_info = get_spotify_info(track_name, session)
|
spotify_info = get_spotify_info(track_name, session)
|
||||||
spotify_info = {} # TODO: add proxy for info retrieve
|
|
||||||
yandex_info = search_yandex(track_name)
|
yandex_info = search_yandex(track_name)
|
||||||
if "album_image_path" in spotify_info and "album_image_path" in yandex_info:
|
if "album_image_path" in spotify_info and "album_image_path" in yandex_info:
|
||||||
os.remove(yandex_info["album_image_path"])
|
os.remove(yandex_info["album_image_path"])
|
||||||
|
|
|
@ -1,156 +1,48 @@
|
||||||
from django.core.cache import cache
|
|
||||||
from django.db.models import Case, When
|
from django.db.models import Case, When
|
||||||
from django_elasticsearch_dsl.registries import registry
|
|
||||||
from elasticsearch_dsl import Q as ES_Q
|
from elasticsearch_dsl import Q as ES_Q
|
||||||
|
|
||||||
from akarpov.music.documents import AlbumDocument, AuthorDocument, SongDocument
|
from akarpov.music.documents import SongDocument
|
||||||
from akarpov.music.models import Album, Author, Song
|
from akarpov.music.models import Song
|
||||||
|
|
||||||
|
|
||||||
def search_song(query):
|
def search_song(query):
|
||||||
search = SongDocument.search()
|
search = SongDocument.search()
|
||||||
|
search_query = ES_Q(
|
||||||
should_queries = [
|
"bool",
|
||||||
ES_Q("match_phrase", name={"query": query, "boost": 5}),
|
should=[
|
||||||
ES_Q(
|
ES_Q(
|
||||||
"nested",
|
"multi_match",
|
||||||
path="authors",
|
query=query,
|
||||||
query=ES_Q("match_phrase", name={"query": query, "boost": 4}),
|
fields=["name^5", "authors.name^3", "album.name^3"],
|
||||||
),
|
fuzziness="AUTO",
|
||||||
ES_Q(
|
|
||||||
"nested",
|
|
||||||
path="album",
|
|
||||||
query=ES_Q("match_phrase", name={"query": query, "boost": 4}),
|
|
||||||
),
|
|
||||||
ES_Q("match", name={"query": query, "fuzziness": "AUTO", "boost": 3}),
|
|
||||||
ES_Q(
|
|
||||||
"nested",
|
|
||||||
path="authors",
|
|
||||||
query=ES_Q("match", name={"query": query, "fuzziness": "AUTO", "boost": 2}),
|
|
||||||
),
|
|
||||||
ES_Q(
|
|
||||||
"nested",
|
|
||||||
path="album",
|
|
||||||
query=ES_Q("match", name={"query": query, "fuzziness": "AUTO", "boost": 2}),
|
|
||||||
),
|
|
||||||
ES_Q("wildcard", name={"value": f"*{query.lower()}*", "boost": 1}),
|
|
||||||
ES_Q(
|
|
||||||
"nested",
|
|
||||||
path="authors",
|
|
||||||
query=ES_Q("wildcard", name={"value": f"*{query.lower()}*", "boost": 0.8}),
|
|
||||||
),
|
|
||||||
ES_Q(
|
|
||||||
"nested",
|
|
||||||
path="album",
|
|
||||||
query=ES_Q("wildcard", name={"value": f"*{query.lower()}*", "boost": 0.8}),
|
|
||||||
),
|
|
||||||
ES_Q(
|
|
||||||
"match",
|
|
||||||
name_transliterated={"query": query, "fuzziness": "AUTO", "boost": 1},
|
|
||||||
),
|
|
||||||
ES_Q(
|
|
||||||
"nested",
|
|
||||||
path="authors",
|
|
||||||
query=ES_Q(
|
|
||||||
"match",
|
|
||||||
name_transliterated={"query": query, "fuzziness": "AUTO", "boost": 0.8},
|
|
||||||
),
|
),
|
||||||
),
|
ES_Q("wildcard", name__raw=f"*{query.lower()}*"),
|
||||||
ES_Q(
|
ES_Q(
|
||||||
"nested",
|
"nested",
|
||||||
path="album",
|
path="authors",
|
||||||
query=ES_Q(
|
query=ES_Q("wildcard", authors__name__raw=f"*{query.lower()}*"),
|
||||||
"match",
|
|
||||||
name_transliterated={"query": query, "fuzziness": "AUTO", "boost": 0.8},
|
|
||||||
),
|
),
|
||||||
),
|
ES_Q(
|
||||||
]
|
"nested",
|
||||||
|
path="album",
|
||||||
|
query=ES_Q("wildcard", album__name__raw=f"*{query.lower()}*"),
|
||||||
|
),
|
||||||
|
ES_Q("wildcard", meta__raw=f"*{query.lower()}*"),
|
||||||
|
],
|
||||||
|
minimum_should_match=1,
|
||||||
|
)
|
||||||
|
|
||||||
search_query = ES_Q("bool", should=should_queries, minimum_should_match=1)
|
search = search.query(search_query)
|
||||||
|
|
||||||
search = search.query(search_query).extra(size=20)
|
|
||||||
response = search.execute()
|
response = search.execute()
|
||||||
|
|
||||||
|
# Check for hits and get song instances
|
||||||
if response.hits:
|
if response.hits:
|
||||||
hit_ids = [hit.meta.id for hit in response.hits]
|
hit_ids = [hit.meta.id for hit in response.hits]
|
||||||
songs = Song.objects.filter(id__in=hit_ids).order_by(
|
songs = Song.objects.filter(id__in=hit_ids).order_by(
|
||||||
Case(*[When(pk=pk, then=pos) for pos, pk in enumerate(hit_ids)])
|
Case(*[When(pk=pk, then=pos) for pos, pk in enumerate(hit_ids)])
|
||||||
)
|
)
|
||||||
|
|
||||||
return songs
|
return songs
|
||||||
|
|
||||||
return Song.objects.none()
|
return Song.objects.none()
|
||||||
|
|
||||||
|
|
||||||
def autocomplete_search(query):
|
|
||||||
s = SongDocument.search()
|
|
||||||
s = s.suggest("song_suggest", query, completion={"field": "suggest"})
|
|
||||||
suggestions = s.execute().suggest.song_suggest[0].options
|
|
||||||
return [option.text for option in suggestions]
|
|
||||||
|
|
||||||
|
|
||||||
def get_popular_songs():
|
|
||||||
if "popular_songs" in cache:
|
|
||||||
return cache.get("popular_songs")
|
|
||||||
else:
|
|
||||||
songs = Song.objects.filter(played__gt=300).order_by("-played")[:10]
|
|
||||||
cache.set("popular_songs", songs, timeout=3600)
|
|
||||||
return songs
|
|
||||||
|
|
||||||
|
|
||||||
def bulk_update_index(model_class):
|
|
||||||
qs = model_class.objects.all()
|
|
||||||
registry.update(qs, bulk_size=100)
|
|
||||||
|
|
||||||
|
|
||||||
def search_author(query):
|
|
||||||
search = AuthorDocument.search()
|
|
||||||
|
|
||||||
should_queries = [
|
|
||||||
ES_Q("match_phrase", name={"query": query, "boost": 5}),
|
|
||||||
ES_Q("match", name={"query": query, "fuzziness": "AUTO", "boost": 3}),
|
|
||||||
ES_Q("wildcard", name={"value": f"*{query.lower()}*", "boost": 1}),
|
|
||||||
ES_Q(
|
|
||||||
"match",
|
|
||||||
name_transliterated={"query": query, "fuzziness": "AUTO", "boost": 1},
|
|
||||||
),
|
|
||||||
]
|
|
||||||
|
|
||||||
search_query = ES_Q("bool", should=should_queries, minimum_should_match=1)
|
|
||||||
search = search.query(search_query).extra(size=10)
|
|
||||||
response = search.execute()
|
|
||||||
|
|
||||||
if response.hits:
|
|
||||||
hit_ids = [hit.meta.id for hit in response.hits]
|
|
||||||
authors = Author.objects.filter(id__in=hit_ids).order_by(
|
|
||||||
Case(*[When(pk=pk, then=pos) for pos, pk in enumerate(hit_ids)])
|
|
||||||
)
|
|
||||||
return authors
|
|
||||||
|
|
||||||
return Author.objects.none()
|
|
||||||
|
|
||||||
|
|
||||||
def search_album(query):
|
|
||||||
search = AlbumDocument.search()
|
|
||||||
|
|
||||||
should_queries = [
|
|
||||||
ES_Q("match_phrase", name={"query": query, "boost": 5}),
|
|
||||||
ES_Q("match", name={"query": query, "fuzziness": "AUTO", "boost": 3}),
|
|
||||||
ES_Q("wildcard", name={"value": f"*{query.lower()}*", "boost": 1}),
|
|
||||||
ES_Q(
|
|
||||||
"match",
|
|
||||||
name_transliterated={"query": query, "fuzziness": "AUTO", "boost": 1},
|
|
||||||
),
|
|
||||||
]
|
|
||||||
|
|
||||||
search_query = ES_Q("bool", should=should_queries, minimum_should_match=1)
|
|
||||||
search = search.query(search_query).extra(size=10)
|
|
||||||
response = search.execute()
|
|
||||||
|
|
||||||
if response.hits:
|
|
||||||
hit_ids = [hit.meta.id for hit in response.hits]
|
|
||||||
albums = Album.objects.filter(id__in=hit_ids).order_by(
|
|
||||||
Case(*[When(pk=pk, then=pos) for pos, pk in enumerate(hit_ids)])
|
|
||||||
)
|
|
||||||
return albums
|
|
||||||
|
|
||||||
return Album.objects.none()
|
|
||||||
|
|
|
@ -75,40 +75,13 @@ def load_file_meta(track: int, user_id: int) -> str:
|
||||||
return str(song)
|
return str(song)
|
||||||
|
|
||||||
|
|
||||||
def load_url(link: str, user_id: int):
|
def load_playlist(link: str, user_id: int):
|
||||||
|
author = link.split("/")[4]
|
||||||
|
playlist_id = link.split("/")[-1]
|
||||||
|
|
||||||
client = login()
|
client = login()
|
||||||
obj_id = link.split("/")[-1]
|
playlist = client.users_playlists(int(playlist_id), author) # type: Playlist
|
||||||
obj_id = obj_id.split("?")[0]
|
for track in playlist.fetch_tracks():
|
||||||
try:
|
|
||||||
obj_id = int(obj_id)
|
|
||||||
except ValueError:
|
|
||||||
print("Invalid link")
|
|
||||||
return None
|
|
||||||
|
|
||||||
if "/playlists/" in link:
|
|
||||||
author = link.split("/")[4]
|
|
||||||
|
|
||||||
playlist = client.users_playlists(obj_id, author) # type: Playlist
|
|
||||||
for track in playlist.fetch_tracks():
|
|
||||||
tasks.load_ym_file_meta.apply_async(
|
|
||||||
kwargs={"track": track.track.id, "user_id": user_id}
|
|
||||||
)
|
|
||||||
elif "/album/" in link:
|
|
||||||
album = client.albums_with_tracks(obj_id)
|
|
||||||
for volume in album.volumes:
|
|
||||||
for track in volume:
|
|
||||||
tasks.load_ym_file_meta.apply_async(
|
|
||||||
kwargs={"track": track.id, "user_id": user_id}
|
|
||||||
)
|
|
||||||
elif "/artist/" in link:
|
|
||||||
artist = client.artists(obj_id)[0]
|
|
||||||
albums = artist.get_albums(page_size=100)
|
|
||||||
for album in albums:
|
|
||||||
for track in album.fetch_tracks():
|
|
||||||
tasks.load_ym_file_meta.apply_async(
|
|
||||||
kwargs={"track": track.id, "user_id": user_id}
|
|
||||||
)
|
|
||||||
else:
|
|
||||||
tasks.load_ym_file_meta.apply_async(
|
tasks.load_ym_file_meta.apply_async(
|
||||||
kwargs={"track": obj_id, "user_id": user_id}
|
kwargs={"track": track.track.id, "user_id": user_id}
|
||||||
)
|
)
|
||||||
|
|
|
@ -2,190 +2,103 @@
|
||||||
from urllib.parse import parse_qs, urlparse
|
from urllib.parse import parse_qs, urlparse
|
||||||
|
|
||||||
import pylast
|
import pylast
|
||||||
import requests
|
|
||||||
import spotipy
|
import spotipy
|
||||||
import structlog
|
import structlog
|
||||||
|
import ytmusicapi
|
||||||
from asgiref.sync import async_to_sync
|
from asgiref.sync import async_to_sync
|
||||||
from celery import shared_task
|
from celery import shared_task
|
||||||
from channels.layers import get_channel_layer
|
from channels.layers import get_channel_layer
|
||||||
from django.conf import settings
|
from django.conf import settings
|
||||||
from django.shortcuts import get_object_or_404
|
|
||||||
from django.utils import timezone
|
from django.utils import timezone
|
||||||
from django.utils.timezone import now
|
from django.utils.timezone import now
|
||||||
from spotipy import SpotifyClientCredentials
|
from spotipy import SpotifyClientCredentials
|
||||||
from ytmusicapi import YTMusic
|
|
||||||
|
|
||||||
from akarpov.music.api.serializers import SongSerializer
|
from akarpov.music.api.serializers import SongSerializer
|
||||||
from akarpov.music.models import (
|
from akarpov.music.models import (
|
||||||
AnonMusicUser,
|
AnonMusicUser,
|
||||||
AnonMusicUserHistory,
|
AnonMusicUserHistory,
|
||||||
MusicDraft,
|
|
||||||
RadioSong,
|
RadioSong,
|
||||||
Song,
|
Song,
|
||||||
UserListenHistory,
|
UserListenHistory,
|
||||||
UserMusicProfile,
|
UserMusicProfile,
|
||||||
)
|
)
|
||||||
from akarpov.music.services import spotify, yandex, youtube
|
from akarpov.music.services import spotify, yandex, youtube
|
||||||
from akarpov.music.services.drafts import save_song_from_draft
|
|
||||||
from akarpov.music.services.file import load_dir, load_file
|
from akarpov.music.services.file import load_dir, load_file
|
||||||
from akarpov.utils.celery import get_scheduled_tasks_name
|
from akarpov.utils.celery import get_scheduled_tasks_name
|
||||||
|
|
||||||
logger = structlog.get_logger(__name__)
|
logger = structlog.get_logger(__name__)
|
||||||
|
|
||||||
|
|
||||||
@shared_task(soft_time_limit=60 * 60, time_limit=60 * 120)
|
@shared_task(soft_time_limit=60 * 20, time_limit=60 * 30)
|
||||||
def list_tracks(self, url: str, user_id: int | None = None) -> str | None:
|
def list_tracks(url, user_id):
|
||||||
"""Update list_tracks to handle failures"""
|
if "music.youtube.com" in url or "youtu.be" in url:
|
||||||
try:
|
url = url.replace("music.youtube.com", "youtube.com")
|
||||||
url = normalize_url(url)
|
url = url.replace("youtu.be", "youtube.com")
|
||||||
handlers = {
|
if "spotify.com" in url:
|
||||||
"spotify.com": handle_spotify,
|
spotify.download_url(url, user_id)
|
||||||
"music.yandex.ru": handle_yandex,
|
elif "music.yandex.ru" in url:
|
||||||
"youtube.com": handle_youtube,
|
yandex.load_playlist(url, user_id)
|
||||||
}
|
if "youtube.com" in url:
|
||||||
|
if "channel" in url or "/c/" in url:
|
||||||
|
ytmusic = ytmusicapi.YTMusic()
|
||||||
|
channel_id = url.split("/")[-1]
|
||||||
|
channel_songs = ytmusic.get_artist(channel_id)["songs"]["results"]
|
||||||
|
|
||||||
for domain, handler in handlers.items():
|
for song in channel_songs:
|
||||||
if domain in url:
|
process_yb.apply_async(
|
||||||
return handler(url, user_id)
|
kwargs={
|
||||||
|
"url": f"https://youtube.com/watch?v={song['videoId']}",
|
||||||
|
"user_id": user_id,
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
return fallback_search(url, user_id)
|
elif "playlist" in url or "&list=" in url:
|
||||||
except Exception as e:
|
ytmusic = ytmusicapi.YTMusic()
|
||||||
draft = MusicDraft.objects.create(
|
|
||||||
provider="unknown",
|
|
||||||
original_url=url,
|
|
||||||
user_id=user_id,
|
|
||||||
status="pending",
|
|
||||||
error_message=str(e),
|
|
||||||
)
|
|
||||||
handle_download_failure.delay(str(draft.id), self.request.id)
|
|
||||||
return None
|
|
||||||
|
|
||||||
|
# Parse the URL and the query string
|
||||||
|
parsed_url = urlparse(url)
|
||||||
|
parsed_qs = parse_qs(parsed_url.query)
|
||||||
|
|
||||||
def normalize_url(url):
|
# Get the playlist ID from the parsed query string
|
||||||
return url.replace("music.youtube.com", "youtube.com").replace(
|
playlist_id = parsed_qs.get("list", [None])[0]
|
||||||
"youtu.be", "youtube.com"
|
|
||||||
)
|
|
||||||
|
|
||||||
|
if playlist_id:
|
||||||
|
playlist_songs = ytmusic.get_playlist(playlist_id)["tracks"]
|
||||||
|
|
||||||
def handle_spotify(url: str, user_id: int | None = None) -> str | None:
|
else:
|
||||||
return download_spotify_url.delay(url, user_id)
|
raise ValueError("No playlist ID found in the URL.")
|
||||||
|
for song in playlist_songs:
|
||||||
|
process_yb.apply_async(
|
||||||
def handle_yandex(url: str, user_id: int | None = None) -> str | None:
|
kwargs={
|
||||||
return load_yandex_url.delay(url, user_id)
|
"url": f"https://music.youtube.com/watch?v={song['videoId']}",
|
||||||
|
"user_id": user_id,
|
||||||
|
}
|
||||||
def handle_youtube(url: str, user_id: int | None = None) -> str | None:
|
)
|
||||||
"""Handle YouTube downloads"""
|
else:
|
||||||
if "channel" in url or "/c/" in url:
|
process_yb.apply_async(kwargs={"url": url, "user_id": user_id})
|
||||||
return handle_youtube_channel(url, user_id)
|
|
||||||
elif "playlist" in url or "&list=" in url:
|
|
||||||
return handle_youtube_playlist(url, user_id)
|
|
||||||
else:
|
else:
|
||||||
return process_yb.delay(url, user_id)
|
spotify_manager = SpotifyClientCredentials(
|
||||||
|
client_id=settings.MUSIC_SPOTIFY_ID,
|
||||||
|
client_secret=settings.MUSIC_SPOTIFY_SECRET,
|
||||||
def handle_youtube_channel(url, user_id):
|
|
||||||
ytmusic = YTMusic()
|
|
||||||
channel_id = url.split("/")[-1]
|
|
||||||
channel_songs = ytmusic.get_artist(channel_id)["songs"]["results"]
|
|
||||||
|
|
||||||
for song in channel_songs:
|
|
||||||
process_yb.apply_async(
|
|
||||||
kwargs={
|
|
||||||
"url": f"https://youtube.com/watch?v={song['videoId']}",
|
|
||||||
"user_id": user_id,
|
|
||||||
}
|
|
||||||
)
|
)
|
||||||
return url
|
spotify_search = spotipy.Spotify(client_credentials_manager=spotify_manager)
|
||||||
|
|
||||||
|
results = spotify_search.search(q=url, type="track", limit=1)
|
||||||
def handle_youtube_playlist(url, user_id):
|
top_track = (
|
||||||
ytmusic = YTMusic()
|
results["tracks"]["items"][0] if results["tracks"]["items"] else None
|
||||||
parsed_url = urlparse(url)
|
|
||||||
parsed_qs = parse_qs(parsed_url.query)
|
|
||||||
playlist_id = parsed_qs.get("list", [None])[0]
|
|
||||||
|
|
||||||
if not playlist_id:
|
|
||||||
raise ValueError("No playlist ID found in the URL.")
|
|
||||||
|
|
||||||
playlist_songs = ytmusic.get_playlist(playlist_id)["tracks"]
|
|
||||||
|
|
||||||
for song in playlist_songs:
|
|
||||||
process_yb.apply_async(
|
|
||||||
kwargs={
|
|
||||||
"url": f"https://music.youtube.com/watch?v={song['videoId']}",
|
|
||||||
"user_id": user_id,
|
|
||||||
}
|
|
||||||
)
|
)
|
||||||
return url
|
|
||||||
|
|
||||||
|
if top_track:
|
||||||
def fallback_search(url, user_id):
|
spotify.download_url(top_track["external_urls"]["spotify"], user_id)
|
||||||
spotify_manager = SpotifyClientCredentials(
|
url = top_track["external_urls"]["spotify"]
|
||||||
client_id=settings.MUSIC_SPOTIFY_ID,
|
|
||||||
client_secret=settings.MUSIC_SPOTIFY_SECRET,
|
|
||||||
)
|
|
||||||
spotify_search = spotipy.Spotify(client_credentials_manager=spotify_manager)
|
|
||||||
|
|
||||||
results = spotify_search.search(q=url, type="track", limit=1)
|
|
||||||
top_track = results["tracks"]["items"][0] if results["tracks"]["items"] else None
|
|
||||||
|
|
||||||
if top_track:
|
|
||||||
spotify_url = top_track["external_urls"]["spotify"]
|
|
||||||
spotify.download_url(spotify_url, user_id)
|
|
||||||
return spotify_url
|
|
||||||
|
|
||||||
return url
|
return url
|
||||||
|
|
||||||
|
|
||||||
@shared_task(bind=True)
|
@shared_task(max_retries=5)
|
||||||
def process_yb(self, url: str, user_id: int | None = None) -> str | None:
|
def process_yb(url, user_id):
|
||||||
"""Update YouTube download to handle failures"""
|
youtube.download_from_youtube_link(url, user_id)
|
||||||
try:
|
return url
|
||||||
return str(youtube.download_from_youtube_link(url, user_id))
|
|
||||||
except Exception as e:
|
|
||||||
draft = MusicDraft.objects.create(
|
|
||||||
provider="youtube",
|
|
||||||
original_url=url,
|
|
||||||
user_id=user_id,
|
|
||||||
status="pending",
|
|
||||||
error_message=str(e),
|
|
||||||
)
|
|
||||||
handle_download_failure.delay(str(draft.id), self.request.id)
|
|
||||||
return None
|
|
||||||
|
|
||||||
|
|
||||||
@shared_task(bind=True)
|
|
||||||
def download_spotify_url(self, url: str, user_id: int | None = None) -> str | None:
|
|
||||||
try:
|
|
||||||
return spotify.download_url(url, user_id)
|
|
||||||
except Exception as e:
|
|
||||||
draft = MusicDraft.objects.create(
|
|
||||||
provider="spotify",
|
|
||||||
original_url=url,
|
|
||||||
user_id=user_id,
|
|
||||||
status="pending",
|
|
||||||
error_message=str(e),
|
|
||||||
)
|
|
||||||
handle_download_failure.delay(str(draft.id), self.request.id)
|
|
||||||
return None
|
|
||||||
|
|
||||||
|
|
||||||
@shared_task(bind=True)
|
|
||||||
def load_yandex_url(self, url: str, user_id: int | None = None) -> str | None:
|
|
||||||
try:
|
|
||||||
return yandex.load_url(url, user_id)
|
|
||||||
except Exception as e:
|
|
||||||
draft = MusicDraft.objects.create(
|
|
||||||
provider="yandex",
|
|
||||||
original_url=url,
|
|
||||||
user_id=user_id,
|
|
||||||
status="pending",
|
|
||||||
error_message=str(e),
|
|
||||||
)
|
|
||||||
handle_download_failure.delay(str(draft.id), self.request.id)
|
|
||||||
return None
|
|
||||||
|
|
||||||
|
|
||||||
@shared_task
|
@shared_task
|
||||||
|
@ -298,71 +211,14 @@ def listen_to_song(song_id, user_id=None, anon=True):
|
||||||
session_key=lastfm_token,
|
session_key=lastfm_token,
|
||||||
)
|
)
|
||||||
song = Song.objects.get(id=song_id)
|
song = Song.objects.get(id=song_id)
|
||||||
artist_name = song.get_first_author_name()
|
artist_name = song.artists_names
|
||||||
track_name = song.name
|
track_name = song.name
|
||||||
album_name = song.album.name
|
|
||||||
timestamp = int(timezone.now().timestamp())
|
timestamp = int(timezone.now().timestamp())
|
||||||
network.scrobble(
|
network.scrobble(
|
||||||
artist=artist_name,
|
artist=artist_name, title=track_name, timestamp=timestamp
|
||||||
title=track_name,
|
|
||||||
timestamp=timestamp,
|
|
||||||
album=album_name,
|
|
||||||
)
|
|
||||||
network.update_now_playing(
|
|
||||||
artist=artist_name, title=track_name, album=album_name
|
|
||||||
)
|
)
|
||||||
except UserMusicProfile.DoesNotExist:
|
except UserMusicProfile.DoesNotExist:
|
||||||
pass
|
pass
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.error(f"Last.fm scrobble error: {e}")
|
logger.error(f"Last.fm scrobble error: {e}")
|
||||||
return song_id
|
return song_id
|
||||||
|
|
||||||
|
|
||||||
@shared_task
|
|
||||||
def handle_download_failure(draft_id: str, original_task_id: str):
|
|
||||||
"""
|
|
||||||
Handle failed downloads by sending request to external service
|
|
||||||
"""
|
|
||||||
draft = MusicDraft.objects.get(id=draft_id)
|
|
||||||
external_service_url = "http://music-download-service/api/v1/download"
|
|
||||||
|
|
||||||
response = requests.post(
|
|
||||||
external_service_url,
|
|
||||||
json={
|
|
||||||
"url": draft.original_url,
|
|
||||||
"provider": draft.provider,
|
|
||||||
"callback_url": draft.get_callback_url(),
|
|
||||||
"file_token": draft.file_token,
|
|
||||||
},
|
|
||||||
)
|
|
||||||
|
|
||||||
if response.status_code != 202:
|
|
||||||
draft.status = "failed"
|
|
||||||
draft.error_message = f"External service request failed: {response.text}"
|
|
||||||
draft.save()
|
|
||||||
return False
|
|
||||||
|
|
||||||
draft.status = "processing"
|
|
||||||
draft.save()
|
|
||||||
return True
|
|
||||||
|
|
||||||
|
|
||||||
@shared_task
|
|
||||||
def process_draft_callback(
|
|
||||||
draft_id: str, status: str, meta_data: dict = None, error_message: str = None
|
|
||||||
):
|
|
||||||
"""
|
|
||||||
Process callback from external service
|
|
||||||
"""
|
|
||||||
draft = get_object_or_404(MusicDraft, id=draft_id)
|
|
||||||
draft.status = status
|
|
||||||
|
|
||||||
if meta_data:
|
|
||||||
draft.meta_data = meta_data
|
|
||||||
if error_message:
|
|
||||||
draft.error_message = error_message
|
|
||||||
|
|
||||||
draft.save()
|
|
||||||
|
|
||||||
if status == "complete":
|
|
||||||
save_song_from_draft(draft)
|
|
||||||
|
|
|
@ -2,7 +2,7 @@
|
||||||
|
|
||||||
{% block content %}
|
{% block content %}
|
||||||
<h1>Welcome to music app</h1>
|
<h1>Welcome to music app</h1>
|
||||||
<p>This is mainly the backend of music, you should consider using side clients like: <a href="https://next.akarpov.ru/music">otomir23's client</a> or my <a href="https://t.me/akarpov_music_bot">inline telegram bot</a></p>
|
<p>This is mainly the backend of music, you should consider using side clients like: <a href="https://next.akarpov.ru/music">otomir23's client</a></p>
|
||||||
{% if request.user.is_authenticated %}
|
{% if request.user.is_authenticated %}
|
||||||
{% if last_fm_account %}
|
{% if last_fm_account %}
|
||||||
<p>Last.fm connected to {{ last_fm_account }}, <a href="{% url 'music:lastfm_connect' %}">reconnect</a></p>
|
<p>Last.fm connected to {{ last_fm_account }}, <a href="{% url 'music:lastfm_connect' %}">reconnect</a></p>
|
||||||
|
|
|
@ -1,8 +1,6 @@
|
||||||
from drf_spectacular.extensions import OpenApiAuthenticationExtension
|
|
||||||
from drf_spectacular.plumbing import build_bearer_security_scheme_object
|
|
||||||
from rest_framework.authentication import BaseAuthentication
|
from rest_framework.authentication import BaseAuthentication
|
||||||
|
|
||||||
from akarpov.users.models import User, UserAPIToken
|
from akarpov.users.models import UserAPIToken
|
||||||
from akarpov.users.tasks import set_last_active_token
|
from akarpov.users.tasks import set_last_active_token
|
||||||
|
|
||||||
|
|
||||||
|
@ -21,14 +19,4 @@ def authenticate(self, request):
|
||||||
return None
|
return None
|
||||||
set_last_active_token.delay(token.token)
|
set_last_active_token.delay(token.token)
|
||||||
|
|
||||||
return User.objects.cache().get(id=token.user_id), token
|
return token.user, token
|
||||||
|
|
||||||
|
|
||||||
class UserTokenAuthenticationExtension(OpenApiAuthenticationExtension):
|
|
||||||
target_class = "akarpov.users.api.authentification.UserTokenAuthentication"
|
|
||||||
name = "UserTokenAuthentication"
|
|
||||||
|
|
||||||
def get_security_definition(self, auto_schema):
|
|
||||||
return build_bearer_security_scheme_object(
|
|
||||||
header_name="Authorization", token_prefix="Bearer"
|
|
||||||
)
|
|
||||||
|
|
|
@ -18,8 +18,6 @@
|
||||||
)
|
)
|
||||||
from akarpov.users.models import User
|
from akarpov.users.models import User
|
||||||
|
|
||||||
from .authentification import UserTokenAuthentication # noqa: F401
|
|
||||||
|
|
||||||
|
|
||||||
class UserRegisterAPIViewSet(generics.CreateAPIView):
|
class UserRegisterAPIViewSet(generics.CreateAPIView):
|
||||||
"""Creates new user and sends verification email"""
|
"""Creates new user and sends verification email"""
|
||||||
|
|
|
@ -214,16 +214,23 @@ def list_tokens(request):
|
||||||
@login_required
|
@login_required
|
||||||
def create_token(request):
|
def create_token(request):
|
||||||
initial_data = {}
|
initial_data = {}
|
||||||
|
|
||||||
|
# Обработка параметров 'name' и 'active_until'
|
||||||
if "name" in request.GET:
|
if "name" in request.GET:
|
||||||
initial_data["name"] = request.GET["name"]
|
initial_data["name"] = request.GET["name"]
|
||||||
if "active_until" in request.GET:
|
if "active_until" in request.GET:
|
||||||
initial_data["active_until"] = request.GET["active_until"]
|
initial_data["active_until"] = request.GET["active_until"]
|
||||||
|
|
||||||
|
# Создаем QueryDict для разрешений, чтобы правильно обработать повторяющиеся ключи
|
||||||
permissions_query_dict = QueryDict("", mutable=True)
|
permissions_query_dict = QueryDict("", mutable=True)
|
||||||
|
|
||||||
|
# Разбор параметров разрешений
|
||||||
permissions = request.GET.getlist("permissions")
|
permissions = request.GET.getlist("permissions")
|
||||||
for perm in permissions:
|
for perm in permissions:
|
||||||
category, permission = perm.split(".")
|
category, permission = perm.split(".")
|
||||||
permissions_query_dict.update({f"permissions_{category}": [permission]})
|
permissions_query_dict.update({f"permissions_{category}": [permission]})
|
||||||
|
|
||||||
|
# Переводим QueryDict в обычный словарь для использования в initial
|
||||||
permissions_data = {key: value for key, value in permissions_query_dict.lists()}
|
permissions_data = {key: value for key, value in permissions_query_dict.lists()}
|
||||||
|
|
||||||
initial_data.update(permissions_data)
|
initial_data.update(permissions_data)
|
||||||
|
@ -235,6 +242,7 @@ def create_token(request):
|
||||||
initial=initial_data, permissions_context=UserAPIToken.permission_template
|
initial=initial_data, permissions_context=UserAPIToken.permission_template
|
||||||
)
|
)
|
||||||
if request.method == "POST":
|
if request.method == "POST":
|
||||||
|
print(request.POST)
|
||||||
form = TokenCreationForm(request.POST)
|
form = TokenCreationForm(request.POST)
|
||||||
if form.is_valid():
|
if form.is_valid():
|
||||||
new_token = form.save(commit=False)
|
new_token = form.save(commit=False)
|
||||||
|
|
|
@ -28,6 +28,8 @@ RUN apt-get update && \
|
||||||
apt-get install -y build-essential libpq-dev gettext libmagic-dev libjpeg-dev zlib1g-dev && \
|
apt-get install -y build-essential libpq-dev gettext libmagic-dev libjpeg-dev zlib1g-dev && \
|
||||||
# Dependencies for file preview generation
|
# Dependencies for file preview generation
|
||||||
apt-get install -y webp git libimage-exiftool-perl libmagickwand-dev ffmpeg libgdal-dev && \
|
apt-get install -y webp git libimage-exiftool-perl libmagickwand-dev ffmpeg libgdal-dev && \
|
||||||
|
# ML dependencies \
|
||||||
|
# none for now
|
||||||
apt-get purge -y --auto-remove -o APT:AutoRemove:RecommendsImportant=false && \
|
apt-get purge -y --auto-remove -o APT:AutoRemove:RecommendsImportant=false && \
|
||||||
rm -rf /var/lib/apt/lists/*
|
rm -rf /var/lib/apt/lists/*
|
||||||
|
|
||||||
|
@ -46,6 +48,7 @@ RUN poetry export --without-hashes -f requirements.txt | /venv/bin/pip install -
|
||||||
|
|
||||||
COPY . .
|
COPY . .
|
||||||
RUN poetry build && /venv/bin/pip install dist/*.whl
|
RUN poetry build && /venv/bin/pip install dist/*.whl
|
||||||
|
RUN /venv/bin/python -m nltk.downloader punkt stopwords wordnet
|
||||||
|
|
||||||
|
|
||||||
COPY ./compose/production/django/entrypoint /entrypoint
|
COPY ./compose/production/django/entrypoint /entrypoint
|
||||||
|
@ -68,6 +71,10 @@ COPY ./compose/local/django/start-redirect /start-redirect
|
||||||
RUN sed -i 's/\r$//g' /start-redirect
|
RUN sed -i 's/\r$//g' /start-redirect
|
||||||
RUN chmod +x /start-redirect
|
RUN chmod +x /start-redirect
|
||||||
|
|
||||||
|
COPY ./compose/local/django/install_preview_dependencies /install_preview_dependencies
|
||||||
|
RUN sed -i 's/\r$//g' /install_preview_dependencies
|
||||||
|
RUN chmod +x /install_preview_dependencies
|
||||||
|
|
||||||
COPY ./compose/local/django/celery/worker/start /start-celeryworker
|
COPY ./compose/local/django/celery/worker/start /start-celeryworker
|
||||||
RUN sed -i 's/\r$//g' /start-celeryworker
|
RUN sed -i 's/\r$//g' /start-celeryworker
|
||||||
RUN chmod +x /start-celeryworker
|
RUN chmod +x /start-celeryworker
|
||||||
|
|
|
@ -3,4 +3,6 @@
|
||||||
set -o errexit
|
set -o errexit
|
||||||
set -o nounset
|
set -o nounset
|
||||||
|
|
||||||
|
/install_preview_dependencies
|
||||||
|
|
||||||
celery -A config.celery_app worker --autoscale 20 -l INFO
|
celery -A config.celery_app worker --autoscale 20 -l INFO
|
||||||
|
|
14
compose/local/django/install_preview_dependencies
Normal file
14
compose/local/django/install_preview_dependencies
Normal file
|
@ -0,0 +1,14 @@
|
||||||
|
#!/bin/bash
|
||||||
|
|
||||||
|
apt-get update
|
||||||
|
apt-get install wget libnotify4 scribus libappindicator3-1 libayatana-indicator3-7 libdbusmenu-glib4 libdbusmenu-gtk3-4
|
||||||
|
apt-get install -y poppler-utils libfile-mimeinfo-perl ghostscript libsecret-1-0 zlib1g-dev libjpeg-dev imagemagick libmagic1 libreoffice inkscape xvfb
|
||||||
|
apt-get install -y libxml2-dev libxslt1-dev antiword unrtf tesseract-ocr flac lame libmad0 libsox-fmt-mp3 sox swig
|
||||||
|
apt-get install -y python-dev-is-python3 libxml2-dev libxslt1-dev antiword unrtf poppler-utils tesseract-ocr \
|
||||||
|
flac ffmpeg lame libmad0 libsox-fmt-mp3 sox libjpeg-dev swig
|
||||||
|
wget https://github.com/jgraph/drawio-desktop/releases/download/v13.0.3/draw.io-amd64-13.0.3.deb
|
||||||
|
dpkg -i draw.io-amd64-13.0.3.deb
|
||||||
|
rm draw.io-amd64-13.0.3.deb
|
||||||
|
apt-get purge -y --auto-remove -o APT:AutoRemove:RecommendsImportant=false && \
|
||||||
|
rm -rf /var/lib/apt/lists/*
|
||||||
|
preview --check-dependencies
|
|
@ -80,7 +80,6 @@
|
||||||
"music.*": {"ops": ("fetch", "get", "list"), "timeout": 60 * 15},
|
"music.*": {"ops": ("fetch", "get", "list"), "timeout": 60 * 15},
|
||||||
"otp_totp.totpdevice": {"ops": "all", "timeout": 15 * 60},
|
"otp_totp.totpdevice": {"ops": "all", "timeout": 15 * 60},
|
||||||
"users.userapitoken": {"ops": "all", "timeout": 20 * 60},
|
"users.userapitoken": {"ops": "all", "timeout": 20 * 60},
|
||||||
"users.user": {"ops": "all", "timeout": 5 * 60},
|
|
||||||
}
|
}
|
||||||
CACHEOPS_REDIS = env.str("REDIS_URL")
|
CACHEOPS_REDIS = env.str("REDIS_URL")
|
||||||
|
|
||||||
|
@ -529,11 +528,6 @@
|
||||||
{"url": "http://127.0.0.1:8000", "description": "Local Development server"},
|
{"url": "http://127.0.0.1:8000", "description": "Local Development server"},
|
||||||
{"url": "https://new.akarpov.ru", "description": "Production server"},
|
{"url": "https://new.akarpov.ru", "description": "Production server"},
|
||||||
],
|
],
|
||||||
"EXTENSIONS": {
|
|
||||||
"authentication": [
|
|
||||||
"akarpov.users.api.authentification.UserTokenAuthenticationExtension"
|
|
||||||
],
|
|
||||||
},
|
|
||||||
}
|
}
|
||||||
|
|
||||||
# CKEDITOR
|
# CKEDITOR
|
||||||
|
@ -754,13 +748,6 @@
|
||||||
ELASTICSEARCH_DSL = {
|
ELASTICSEARCH_DSL = {
|
||||||
"default": {"hosts": env("ELASTIC_SEARCH", default="http://127.0.0.1:9200/")},
|
"default": {"hosts": env("ELASTIC_SEARCH", default="http://127.0.0.1:9200/")},
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
USE_DEBUG_TOOLBAR = False
|
USE_DEBUG_TOOLBAR = False
|
||||||
SECURE_PROXY_SSL_HEADER = ("HTTP_X_FORWARDED_PROTO", "https")
|
|
||||||
USE_X_FORWARDED_HOST = True
|
|
||||||
USE_X_FORWARDED_PORT = True
|
|
||||||
|
|
||||||
|
|
||||||
# PREVIEW
|
|
||||||
# ------------------------------------------------------------------------------
|
|
||||||
PREVIEW_SERVICE_URL = env("PREVIEW_SERVICE_URL", default=None)
|
|
||||||
PREVIEW_SERVICE_API_KEY = env("PREVIEW_SERVICE_API_KEY", default=None)
|
|
||||||
|
|
5903
poetry.lock
generated
5903
poetry.lock
generated
File diff suppressed because it is too large
Load Diff
|
@ -69,6 +69,7 @@ channels = {extras = ["daphne"], version = "^4.0.0"}
|
||||||
django-upload-validator = "^1.1.6"
|
django-upload-validator = "^1.1.6"
|
||||||
markdown = "^3.4.4"
|
markdown = "^3.4.4"
|
||||||
pydotplus = "^2.0.2"
|
pydotplus = "^2.0.2"
|
||||||
|
preview-generator = "^0.29"
|
||||||
uuid = "^1.30"
|
uuid = "^1.30"
|
||||||
mutagen = "^1.46.0"
|
mutagen = "^1.46.0"
|
||||||
pydub = "^0.25.1"
|
pydub = "^0.25.1"
|
||||||
|
@ -99,8 +100,11 @@ pytest-mock = "^3.11.1"
|
||||||
pytest-asyncio = "^0.21.1"
|
pytest-asyncio = "^0.21.1"
|
||||||
pytest-lambda = "^2.2.0"
|
pytest-lambda = "^2.2.0"
|
||||||
pgvector = "^0.2.2"
|
pgvector = "^0.2.2"
|
||||||
|
pycld2 = "^0.41"
|
||||||
uuid6 = "^2023.5.2"
|
uuid6 = "^2023.5.2"
|
||||||
uvicorn = "0.23.2"
|
uvicorn = "0.23.2"
|
||||||
|
nltk = "^3.8.1"
|
||||||
|
pymorphy3 = "^1.2.1"
|
||||||
pymorphy3-dicts-ru = "^2.4.417150.4580142"
|
pymorphy3-dicts-ru = "^2.4.417150.4580142"
|
||||||
fastapi = "0.103.0"
|
fastapi = "0.103.0"
|
||||||
pydantic-settings = "^2.0.3"
|
pydantic-settings = "^2.0.3"
|
||||||
|
@ -114,9 +118,9 @@ spotdl = "^4.2.4"
|
||||||
fuzzywuzzy = "^0.18.0"
|
fuzzywuzzy = "^0.18.0"
|
||||||
python-levenshtein = "^0.23.0"
|
python-levenshtein = "^0.23.0"
|
||||||
pylast = "^5.2.0"
|
pylast = "^5.2.0"
|
||||||
|
textract = {git = "https://github.com/Alexander-D-Karpov/textract.git", branch = "master"}
|
||||||
librosa = "^0.10.1"
|
librosa = "^0.10.1"
|
||||||
django-ckeditor-5 = "^0.2.12"
|
django-ckeditor-5 = "^0.2.12"
|
||||||
chardet = "^5.2.0"
|
|
||||||
|
|
||||||
|
|
||||||
[build-system]
|
[build-system]
|
||||||
|
|
0
search/__init__.py
Normal file
0
search/__init__.py
Normal file
6
search/pipeline.py
Normal file
6
search/pipeline.py
Normal file
|
@ -0,0 +1,6 @@
|
||||||
|
from haystack import Document
|
||||||
|
from milvus_haystack import MilvusDocumentStore
|
||||||
|
|
||||||
|
ds = MilvusDocumentStore()
|
||||||
|
ds.write_documents([Document("Some Content")])
|
||||||
|
ds.get_all_documents()
|
2185
search/poetry.lock
generated
Normal file
2185
search/poetry.lock
generated
Normal file
File diff suppressed because it is too large
Load Diff
18
search/pyproject.toml
Normal file
18
search/pyproject.toml
Normal file
|
@ -0,0 +1,18 @@
|
||||||
|
[tool.poetry]
|
||||||
|
name = "search"
|
||||||
|
version = "0.1.0"
|
||||||
|
description = ""
|
||||||
|
authors = ["Alexander-D-Karpov <alexandr.d.karpov@gmail.com>"]
|
||||||
|
readme = "README.md"
|
||||||
|
|
||||||
|
[tool.poetry.dependencies]
|
||||||
|
python = "^3.11"
|
||||||
|
fastapi = "0.99.1"
|
||||||
|
pydantic = "1.10.13"
|
||||||
|
transformers = {version = "4.34.1", extras = ["torch"]}
|
||||||
|
torch = ">=2.0.0, !=2.0.1, !=2.1.0"
|
||||||
|
farm-haystack = {extras = ["faiss"], version = "^1.21.2"}
|
||||||
|
|
||||||
|
[build-system]
|
||||||
|
requires = ["poetry-core>=1.0.0"]
|
||||||
|
build-backend = "poetry.core.masonry.api"
|
4
spacy_setup.sh
Executable file
4
spacy_setup.sh
Executable file
|
@ -0,0 +1,4 @@
|
||||||
|
#!/bin/bash
|
||||||
|
python -m spacy download en_core_web_lg
|
||||||
|
python -m spacy download xx_sent_ud_sm
|
||||||
|
python -m spacy download ru_core_news_lg
|
Loading…
Reference in New Issue
Block a user