Compare commits
3 Commits
v0.1.83.0-
...
v0.1.84.5-
| Author | SHA1 | Date | |
|---|---|---|---|
| e5d93e3af6 | |||
| 938f6c0e3d | |||
| 4e9ae348b9 |
@@ -1,3 +1,15 @@
|
|||||||
|
## 0.1.84.0-dev - 2026-03-16
|
||||||
|
|
||||||
|
- dev: bump to 0.1.83.5-dev Trakt Weiterschauen via watched/shows, Specials überspringen
|
||||||
|
|
||||||
|
## 0.1.83.5-dev - 2026-03-15
|
||||||
|
|
||||||
|
- dev: SerienStream Suche via /suche?term=, Staffel 0 als Filme, Katalog-Suche entfernt
|
||||||
|
|
||||||
|
## 0.1.83.0-dev - 2026-03-15
|
||||||
|
|
||||||
|
- dev: Trakt Performance, Suchfilter Phrase-Match, Debug-Settings Expert-Level
|
||||||
|
|
||||||
## 0.1.82.5-dev - 2026-03-15
|
## 0.1.82.5-dev - 2026-03-15
|
||||||
|
|
||||||
- dev: Update-Versionsvergleich numerisch korrigiert
|
- dev: Update-Versionsvergleich numerisch korrigiert
|
||||||
|
|||||||
@@ -1,5 +1,5 @@
|
|||||||
<?xml version='1.0' encoding='utf-8'?>
|
<?xml version='1.0' encoding='utf-8'?>
|
||||||
<addon id="plugin.video.viewit" name="ViewIt" version="0.1.83.0-dev" provider-name="ViewIt">
|
<addon id="plugin.video.viewit" name="ViewIt" version="0.1.84.5-dev" provider-name="ViewIt">
|
||||||
<requires>
|
<requires>
|
||||||
<import addon="xbmc.python" version="3.0.0" />
|
<import addon="xbmc.python" version="3.0.0" />
|
||||||
<import addon="script.module.requests" />
|
<import addon="script.module.requests" />
|
||||||
|
|||||||
@@ -370,6 +370,40 @@ class TraktClient:
|
|||||||
return []
|
return []
|
||||||
return self._parse_history_items(payload)
|
return self._parse_history_items(payload)
|
||||||
|
|
||||||
|
def get_watched_shows(self, token: str) -> list[TraktItem]:
|
||||||
|
"""GET /users/me/watched/shows – alle Serien mit zuletzt gesehener Episode."""
|
||||||
|
status, payload = self._get("/users/me/watched/shows", token=token)
|
||||||
|
if status != 200 or not isinstance(payload, list):
|
||||||
|
self._do_log(f"get_watched_shows: status={status}")
|
||||||
|
return []
|
||||||
|
result: list[TraktItem] = []
|
||||||
|
for entry in payload:
|
||||||
|
if not isinstance(entry, dict):
|
||||||
|
continue
|
||||||
|
show = entry.get("show") or {}
|
||||||
|
ids = self._parse_ids((show.get("ids") or {}))
|
||||||
|
title = str(show.get("title", "") or "")
|
||||||
|
year = int(show.get("year", 0) or 0)
|
||||||
|
seasons = entry.get("seasons") or []
|
||||||
|
last_season = 0
|
||||||
|
last_episode = 0
|
||||||
|
for s in seasons:
|
||||||
|
snum = int((s.get("number") or 0))
|
||||||
|
if snum == 0: # Specials überspringen
|
||||||
|
continue
|
||||||
|
for ep in (s.get("episodes") or []):
|
||||||
|
enum = int((ep.get("number") or 0))
|
||||||
|
if snum > last_season or (snum == last_season and enum > last_episode):
|
||||||
|
last_season = snum
|
||||||
|
last_episode = enum
|
||||||
|
if title:
|
||||||
|
result.append(TraktItem(
|
||||||
|
title=title, year=year, media_type="episode",
|
||||||
|
ids=ids, season=last_season, episode=last_episode,
|
||||||
|
))
|
||||||
|
self._do_log(f"get_watched_shows: {len(result)} Serien")
|
||||||
|
return result
|
||||||
|
|
||||||
# -------------------------------------------------------------------
|
# -------------------------------------------------------------------
|
||||||
# Calendar
|
# Calendar
|
||||||
# -------------------------------------------------------------------
|
# -------------------------------------------------------------------
|
||||||
|
|||||||
107
addon/default.py
107
addon/default.py
@@ -2095,8 +2095,17 @@ def _run_async(coro):
|
|||||||
"""Fuehrt eine Coroutine aus, auch wenn Kodi bereits einen Event-Loop hat."""
|
"""Fuehrt eine Coroutine aus, auch wenn Kodi bereits einen Event-Loop hat."""
|
||||||
_ensure_windows_selector_policy()
|
_ensure_windows_selector_policy()
|
||||||
|
|
||||||
def _run_with_asyncio_run():
|
def _run_without_asyncio_run():
|
||||||
return asyncio.run(coro)
|
# asyncio.run() wuerde cancel_all_tasks() aufrufen, was auf Android TV
|
||||||
|
# wegen eines kaputten _weakrefset.py-Builds zu NameError: 'len' fuehrt.
|
||||||
|
loop = asyncio.new_event_loop()
|
||||||
|
try:
|
||||||
|
return loop.run_until_complete(coro)
|
||||||
|
finally:
|
||||||
|
try:
|
||||||
|
loop.close()
|
||||||
|
except Exception:
|
||||||
|
pass
|
||||||
|
|
||||||
try:
|
try:
|
||||||
running_loop = asyncio.get_running_loop()
|
running_loop = asyncio.get_running_loop()
|
||||||
@@ -2109,7 +2118,7 @@ def _run_async(coro):
|
|||||||
|
|
||||||
def _worker() -> None:
|
def _worker() -> None:
|
||||||
try:
|
try:
|
||||||
result_box["value"] = _run_with_asyncio_run()
|
result_box["value"] = _run_without_asyncio_run()
|
||||||
except BaseException as exc: # pragma: no cover - defensive
|
except BaseException as exc: # pragma: no cover - defensive
|
||||||
error_box["error"] = exc
|
error_box["error"] = exc
|
||||||
|
|
||||||
@@ -2120,7 +2129,7 @@ def _run_async(coro):
|
|||||||
raise error_box["error"]
|
raise error_box["error"]
|
||||||
return result_box.get("value")
|
return result_box.get("value")
|
||||||
|
|
||||||
return _run_with_asyncio_run()
|
return _run_without_asyncio_run()
|
||||||
|
|
||||||
|
|
||||||
def _series_url_params(plugin: BasisPlugin, title: str) -> dict[str, str]:
|
def _series_url_params(plugin: BasisPlugin, title: str) -> dict[str, str]:
|
||||||
@@ -4414,13 +4423,15 @@ def _play_episode(
|
|||||||
preferred_setter([selected_hoster])
|
preferred_setter([selected_hoster])
|
||||||
|
|
||||||
try:
|
try:
|
||||||
link = plugin.stream_link_for(title, season, episode)
|
with _busy_dialog("Stream wird gesucht..."):
|
||||||
|
link = plugin.stream_link_for(title, season, episode)
|
||||||
if not link:
|
if not link:
|
||||||
_log("Kein Stream gefunden.", xbmc.LOGWARNING)
|
_log("Kein Stream gefunden.", xbmc.LOGWARNING)
|
||||||
xbmcgui.Dialog().notification("Wiedergabe", "Kein Stream gefunden.", xbmcgui.NOTIFICATION_INFO, 3000)
|
xbmcgui.Dialog().notification("Wiedergabe", "Kein Stream gefunden.", xbmcgui.NOTIFICATION_INFO, 3000)
|
||||||
return
|
return
|
||||||
_log(f"Stream-Link: {link}", xbmc.LOGDEBUG)
|
_log(f"Stream-Link: {link}", xbmc.LOGDEBUG)
|
||||||
final_link = _resolve_stream_with_retry(plugin, link)
|
with _busy_dialog("Stream wird aufgelöst..."):
|
||||||
|
final_link = _resolve_stream_with_retry(plugin, link)
|
||||||
if not final_link:
|
if not final_link:
|
||||||
return
|
return
|
||||||
finally:
|
finally:
|
||||||
@@ -4754,13 +4765,33 @@ def _show_collection_titles_page(plugin_name: str, collection: str, page: int =
|
|||||||
xbmcplugin.endOfDirectory(handle)
|
xbmcplugin.endOfDirectory(handle)
|
||||||
return
|
return
|
||||||
titles = [str(t).strip() for t in titles if t and str(t).strip()]
|
titles = [str(t).strip() for t in titles if t and str(t).strip()]
|
||||||
direct_play = bool(plugin_name.casefold() == "einschalten"
|
|
||||||
and _get_setting_bool("einschalten_enable_playback", default=False))
|
|
||||||
for title in titles:
|
|
||||||
_add_directory_item(handle, title, "play_movie" if direct_play else "seasons",
|
|
||||||
{"plugin": plugin_name, "title": title, **_series_url_params(plugin, title)},
|
|
||||||
is_folder=not direct_play)
|
|
||||||
if titles:
|
if titles:
|
||||||
|
use_source, show_tmdb, prefer_source = _metadata_policy(
|
||||||
|
plugin_name, plugin, allow_tmdb=_tmdb_list_enabled()
|
||||||
|
)
|
||||||
|
plugin_meta = _collect_plugin_metadata(plugin, titles) if use_source else {}
|
||||||
|
show_plot = _get_setting_bool("tmdb_show_plot", default=True)
|
||||||
|
show_art = _get_setting_bool("tmdb_show_art", default=True)
|
||||||
|
tmdb_prefetched: dict[str, tuple[dict[str, str], dict[str, str], list[TmdbCastMember]]] = {}
|
||||||
|
tmdb_titles = list(titles) if show_tmdb else []
|
||||||
|
if show_tmdb and prefer_source and use_source:
|
||||||
|
tmdb_titles = [
|
||||||
|
t for t in titles
|
||||||
|
if _needs_tmdb((plugin_meta.get(t) or ({},))[0], (plugin_meta.get(t) or ({}, {}))[1],
|
||||||
|
want_plot=show_plot, want_art=show_art)
|
||||||
|
]
|
||||||
|
if show_tmdb and tmdb_titles:
|
||||||
|
with _busy_dialog(f"{collection} wird geladen..."):
|
||||||
|
tmdb_prefetched = _tmdb_labels_and_art_bulk(tmdb_titles)
|
||||||
|
for title in titles:
|
||||||
|
tmdb_info, tmdb_art, tmdb_cast = tmdb_prefetched.get(title, ({}, {}, [])) if show_tmdb else ({}, {}, [])
|
||||||
|
meta = plugin_meta.get(title)
|
||||||
|
info_labels, art, cast = _merge_metadata(title, tmdb_info, tmdb_art, tmdb_cast, meta)
|
||||||
|
info_labels = dict(info_labels or {})
|
||||||
|
info_labels.setdefault("mediatype", "tvshow")
|
||||||
|
_add_directory_item(handle, title, "seasons",
|
||||||
|
{"plugin": plugin_name, "title": title, **_series_url_params(plugin, title)},
|
||||||
|
is_folder=True, info_labels=info_labels, art=art, cast=cast)
|
||||||
_add_directory_item(handle, "Naechste Seite", "collection_titles_page",
|
_add_directory_item(handle, "Naechste Seite", "collection_titles_page",
|
||||||
{"plugin": plugin_name, "collection": collection, "page": str(page + 1)}, is_folder=True)
|
{"plugin": plugin_name, "collection": collection, "page": str(page + 1)}, is_folder=True)
|
||||||
xbmcplugin.endOfDirectory(handle)
|
xbmcplugin.endOfDirectory(handle)
|
||||||
@@ -4815,11 +4846,33 @@ def _show_tag_titles_page(plugin_name: str, tag: str, page: int = 1) -> None:
|
|||||||
xbmcplugin.endOfDirectory(handle)
|
xbmcplugin.endOfDirectory(handle)
|
||||||
return
|
return
|
||||||
titles = [str(t).strip() for t in titles if t and str(t).strip()]
|
titles = [str(t).strip() for t in titles if t and str(t).strip()]
|
||||||
for title in titles:
|
|
||||||
_add_directory_item(handle, title, "seasons",
|
|
||||||
{"plugin": plugin_name, "title": title, **_series_url_params(plugin, title)},
|
|
||||||
is_folder=True)
|
|
||||||
if titles:
|
if titles:
|
||||||
|
use_source, show_tmdb, prefer_source = _metadata_policy(
|
||||||
|
plugin_name, plugin, allow_tmdb=_tmdb_list_enabled()
|
||||||
|
)
|
||||||
|
plugin_meta = _collect_plugin_metadata(plugin, titles) if use_source else {}
|
||||||
|
show_plot = _get_setting_bool("tmdb_show_plot", default=True)
|
||||||
|
show_art = _get_setting_bool("tmdb_show_art", default=True)
|
||||||
|
tmdb_prefetched: dict[str, tuple[dict[str, str], dict[str, str], list[TmdbCastMember]]] = {}
|
||||||
|
tmdb_titles = list(titles) if show_tmdb else []
|
||||||
|
if show_tmdb and prefer_source and use_source:
|
||||||
|
tmdb_titles = [
|
||||||
|
t for t in titles
|
||||||
|
if _needs_tmdb((plugin_meta.get(t) or ({},))[0], (plugin_meta.get(t) or ({}, {}))[1],
|
||||||
|
want_plot=show_plot, want_art=show_art)
|
||||||
|
]
|
||||||
|
if show_tmdb and tmdb_titles:
|
||||||
|
with _busy_dialog("Schlagwort-Liste wird geladen..."):
|
||||||
|
tmdb_prefetched = _tmdb_labels_and_art_bulk(tmdb_titles)
|
||||||
|
for title in titles:
|
||||||
|
tmdb_info, tmdb_art, tmdb_cast = tmdb_prefetched.get(title, ({}, {}, [])) if show_tmdb else ({}, {}, [])
|
||||||
|
meta = plugin_meta.get(title)
|
||||||
|
info_labels, art, cast = _merge_metadata(title, tmdb_info, tmdb_art, tmdb_cast, meta)
|
||||||
|
info_labels = dict(info_labels or {})
|
||||||
|
info_labels.setdefault("mediatype", "tvshow")
|
||||||
|
_add_directory_item(handle, title, "seasons",
|
||||||
|
{"plugin": plugin_name, "title": title, **_series_url_params(plugin, title)},
|
||||||
|
is_folder=True, info_labels=info_labels, art=art, cast=cast)
|
||||||
_add_directory_item(handle, "Naechste Seite", "tag_titles_page",
|
_add_directory_item(handle, "Naechste Seite", "tag_titles_page",
|
||||||
{"plugin": plugin_name, "tag": tag, "page": str(page + 1)}, is_folder=True)
|
{"plugin": plugin_name, "tag": tag, "page": str(page + 1)}, is_folder=True)
|
||||||
xbmcplugin.endOfDirectory(handle)
|
xbmcplugin.endOfDirectory(handle)
|
||||||
@@ -4929,7 +4982,7 @@ def _show_trakt_watchlist(media_type: str = "") -> None:
|
|||||||
_add_directory_item(handle, label, "search", {"query": item.title}, is_folder=True, info_labels=info_labels, art=art)
|
_add_directory_item(handle, label, "search", {"query": item.title}, is_folder=True, info_labels=info_labels, art=art)
|
||||||
if not items:
|
if not items:
|
||||||
xbmcgui.Dialog().notification("Trakt", "Watchlist ist leer.", xbmcgui.NOTIFICATION_INFO, 3000)
|
xbmcgui.Dialog().notification("Trakt", "Watchlist ist leer.", xbmcgui.NOTIFICATION_INFO, 3000)
|
||||||
xbmcplugin.endOfDirectory(handle)
|
xbmcplugin.endOfDirectory(handle, cacheToDisc=False)
|
||||||
|
|
||||||
|
|
||||||
def _show_trakt_history(page: int = 1) -> None:
|
def _show_trakt_history(page: int = 1) -> None:
|
||||||
@@ -4999,7 +5052,7 @@ def _show_trakt_history(page: int = 1) -> None:
|
|||||||
_add_directory_item(handle, "Naechste Seite >>", "trakt_history", {"page": str(page + 1)}, is_folder=True)
|
_add_directory_item(handle, "Naechste Seite >>", "trakt_history", {"page": str(page + 1)}, is_folder=True)
|
||||||
if not items and page == 1:
|
if not items and page == 1:
|
||||||
xbmcgui.Dialog().notification("Trakt", "Keine History vorhanden.", xbmcgui.NOTIFICATION_INFO, 3000)
|
xbmcgui.Dialog().notification("Trakt", "Keine History vorhanden.", xbmcgui.NOTIFICATION_INFO, 3000)
|
||||||
xbmcplugin.endOfDirectory(handle)
|
xbmcplugin.endOfDirectory(handle, cacheToDisc=False)
|
||||||
|
|
||||||
|
|
||||||
def _show_trakt_upcoming() -> None:
|
def _show_trakt_upcoming() -> None:
|
||||||
@@ -5110,7 +5163,7 @@ def _show_trakt_upcoming() -> None:
|
|||||||
|
|
||||||
_add_directory_item(handle, label, action, params, is_folder=True, info_labels=info_labels, art=art)
|
_add_directory_item(handle, label, action, params, is_folder=True, info_labels=info_labels, art=art)
|
||||||
|
|
||||||
xbmcplugin.endOfDirectory(handle)
|
xbmcplugin.endOfDirectory(handle, cacheToDisc=False)
|
||||||
|
|
||||||
|
|
||||||
def _show_trakt_continue_watching() -> None:
|
def _show_trakt_continue_watching() -> None:
|
||||||
@@ -5127,21 +5180,17 @@ def _show_trakt_continue_watching() -> None:
|
|||||||
_set_content(handle, "episodes")
|
_set_content(handle, "episodes")
|
||||||
|
|
||||||
try:
|
try:
|
||||||
history = client.get_history(token, media_type="episodes", limit=100)
|
watched = client.get_watched_shows(token)
|
||||||
except Exception as exc:
|
except Exception as exc:
|
||||||
_log(f"Trakt History fehlgeschlagen: {exc}", xbmc.LOGWARNING)
|
_log(f"Trakt Watched fehlgeschlagen: {exc}", xbmc.LOGWARNING)
|
||||||
xbmcgui.Dialog().notification("Trakt", "History konnte nicht geladen werden.", xbmcgui.NOTIFICATION_INFO, 3000)
|
xbmcgui.Dialog().notification("Trakt", "Watched-Liste konnte nicht geladen werden.", xbmcgui.NOTIFICATION_INFO, 3000)
|
||||||
xbmcplugin.endOfDirectory(handle)
|
xbmcplugin.endOfDirectory(handle)
|
||||||
return
|
return
|
||||||
|
|
||||||
# Pro Serie nur den zuletzt gesehenen Eintrag behalten (History ist absteigend sortiert)
|
seen: dict[str, object] = {item.title: item for item in watched if item.title}
|
||||||
seen: dict[str, object] = {}
|
|
||||||
for item in history:
|
|
||||||
if item.title and item.title not in seen:
|
|
||||||
seen[item.title] = item
|
|
||||||
|
|
||||||
if not seen:
|
if not seen:
|
||||||
xbmcgui.Dialog().notification("Trakt", "Keine History vorhanden.", xbmcgui.NOTIFICATION_INFO, 3000)
|
xbmcgui.Dialog().notification("Trakt", "Keine gesehenen Serien vorhanden.", xbmcgui.NOTIFICATION_INFO, 3000)
|
||||||
xbmcplugin.endOfDirectory(handle)
|
xbmcplugin.endOfDirectory(handle)
|
||||||
return
|
return
|
||||||
|
|
||||||
@@ -5169,7 +5218,7 @@ def _show_trakt_continue_watching() -> None:
|
|||||||
_, art, _ = tmdb_prefetched.get(last.title, ({}, {}, []))
|
_, art, _ = tmdb_prefetched.get(last.title, ({}, {}, []))
|
||||||
_add_directory_item(handle, display_label, "search", {"query": last.title}, is_folder=True, info_labels=info_labels, art=art)
|
_add_directory_item(handle, display_label, "search", {"query": last.title}, is_folder=True, info_labels=info_labels, art=art)
|
||||||
|
|
||||||
xbmcplugin.endOfDirectory(handle)
|
xbmcplugin.endOfDirectory(handle, cacheToDisc=False)
|
||||||
|
|
||||||
|
|
||||||
# ---------------------------------------------------------------------------
|
# ---------------------------------------------------------------------------
|
||||||
|
|||||||
@@ -57,7 +57,6 @@ else: # pragma: no cover
|
|||||||
|
|
||||||
|
|
||||||
SETTING_BASE_URL = "serienstream_base_url"
|
SETTING_BASE_URL = "serienstream_base_url"
|
||||||
SETTING_CATALOG_SEARCH = "serienstream_catalog_search"
|
|
||||||
DEFAULT_BASE_URL = "https://s.to"
|
DEFAULT_BASE_URL = "https://s.to"
|
||||||
DEFAULT_PREFERRED_HOSTERS = ["voe"]
|
DEFAULT_PREFERRED_HOSTERS = ["voe"]
|
||||||
DEFAULT_TIMEOUT = 20
|
DEFAULT_TIMEOUT = 20
|
||||||
@@ -80,10 +79,7 @@ HEADERS = {
|
|||||||
SESSION_CACHE_TTL_SECONDS = 300
|
SESSION_CACHE_TTL_SECONDS = 300
|
||||||
SESSION_CACHE_PREFIX = "viewit.serienstream"
|
SESSION_CACHE_PREFIX = "viewit.serienstream"
|
||||||
SESSION_CACHE_MAX_TITLE_URLS = 800
|
SESSION_CACHE_MAX_TITLE_URLS = 800
|
||||||
CATALOG_SEARCH_TTL_SECONDS = 600
|
|
||||||
CATALOG_SEARCH_CACHE_KEY = "catalog_index"
|
|
||||||
GENRE_LIST_PAGE_SIZE = 20
|
GENRE_LIST_PAGE_SIZE = 20
|
||||||
_CATALOG_INDEX_MEMORY: tuple[float, list["SeriesResult"]] = (0.0, [])
|
|
||||||
ProgressCallback = Optional[Callable[[str, int | None], Any]]
|
ProgressCallback = Optional[Callable[[str, int | None], Any]]
|
||||||
|
|
||||||
|
|
||||||
@@ -575,8 +571,8 @@ def _search_series_server(query: str) -> list[SeriesResult]:
|
|||||||
if not query:
|
if not query:
|
||||||
return []
|
return []
|
||||||
base = _get_base_url()
|
base = _get_base_url()
|
||||||
search_url = f"{base}/search?q={quote(query)}"
|
search_url = f"{base}/suche?term={quote(query)}"
|
||||||
alt_url = f"{base}/suche?q={quote(query)}"
|
alt_url = f"{base}/search?term={quote(query)}"
|
||||||
for url in (search_url, alt_url):
|
for url in (search_url, alt_url):
|
||||||
try:
|
try:
|
||||||
body = _get_html_simple(url)
|
body = _get_html_simple(url)
|
||||||
@@ -606,158 +602,30 @@ def _search_series_server(query: str) -> list[SeriesResult]:
|
|||||||
continue
|
continue
|
||||||
seen_urls.add(url_abs)
|
seen_urls.add(url_abs)
|
||||||
results.append(SeriesResult(title=title, description="", url=url_abs))
|
results.append(SeriesResult(title=title, description="", url=url_abs))
|
||||||
|
filtered = [r for r in results if _matches_query(query, title=r.title)]
|
||||||
|
if filtered:
|
||||||
|
return filtered
|
||||||
if results:
|
if results:
|
||||||
return results
|
return results
|
||||||
api_results = _search_series_api(query)
|
|
||||||
if api_results:
|
|
||||||
return api_results
|
|
||||||
return []
|
return []
|
||||||
|
|
||||||
|
|
||||||
def _extract_catalog_index_from_html(body: str, *, progress_callback: ProgressCallback = None) -> list[SeriesResult]:
|
|
||||||
items: list[SeriesResult] = []
|
|
||||||
if not body:
|
|
||||||
return items
|
|
||||||
seen_urls: set[str] = set()
|
|
||||||
item_re = re.compile(
|
|
||||||
r"<li[^>]*class=[\"'][^\"']*series-item[^\"']*[\"'][^>]*>(.*?)</li>",
|
|
||||||
re.IGNORECASE | re.DOTALL,
|
|
||||||
)
|
|
||||||
anchor_re = re.compile(r"<a[^>]+href=[\"']([^\"']+)[\"'][^>]*>(.*?)</a>", re.IGNORECASE | re.DOTALL)
|
|
||||||
data_search_re = re.compile(r"data-search=[\"']([^\"']*)[\"']", re.IGNORECASE)
|
|
||||||
for idx, match in enumerate(item_re.finditer(body), start=1):
|
|
||||||
if idx == 1 or idx % 200 == 0:
|
|
||||||
_emit_progress(progress_callback, f"Katalog parsen {idx}", 62)
|
|
||||||
block = match.group(0)
|
|
||||||
inner = match.group(1) or ""
|
|
||||||
anchor_match = anchor_re.search(inner)
|
|
||||||
if not anchor_match:
|
|
||||||
continue
|
|
||||||
href = (anchor_match.group(1) or "").strip()
|
|
||||||
url = _absolute_url(href)
|
|
||||||
if not url or "/serie/" not in url or "/staffel-" in url or "/episode-" in url:
|
|
||||||
continue
|
|
||||||
if url in seen_urls:
|
|
||||||
continue
|
|
||||||
seen_urls.add(url)
|
|
||||||
title_raw = anchor_match.group(2) or ""
|
|
||||||
title = unescape(re.sub(r"\s+", " ", _strip_tags(title_raw))).strip()
|
|
||||||
if not title:
|
|
||||||
continue
|
|
||||||
search_match = data_search_re.search(block)
|
|
||||||
description = (search_match.group(1) or "").strip() if search_match else ""
|
|
||||||
items.append(SeriesResult(title=title, description=description, url=url))
|
|
||||||
return items
|
|
||||||
|
|
||||||
|
|
||||||
def _catalog_index_from_soup(soup: BeautifulSoupT) -> list[SeriesResult]:
|
|
||||||
items: list[SeriesResult] = []
|
|
||||||
if not soup:
|
|
||||||
return items
|
|
||||||
seen_urls: set[str] = set()
|
|
||||||
for item in soup.select("li.series-item"):
|
|
||||||
anchor = item.find("a", href=True)
|
|
||||||
if not anchor:
|
|
||||||
continue
|
|
||||||
href = (anchor.get("href") or "").strip()
|
|
||||||
url = _absolute_url(href)
|
|
||||||
if not url or "/serie/" not in url or "/staffel-" in url or "/episode-" in url:
|
|
||||||
continue
|
|
||||||
if url in seen_urls:
|
|
||||||
continue
|
|
||||||
seen_urls.add(url)
|
|
||||||
title = (anchor.get_text(" ", strip=True) or "").strip()
|
|
||||||
if not title:
|
|
||||||
continue
|
|
||||||
description = (item.get("data-search") or "").strip()
|
|
||||||
items.append(SeriesResult(title=title, description=description, url=url))
|
|
||||||
return items
|
|
||||||
|
|
||||||
|
|
||||||
def _load_catalog_index_from_cache() -> Optional[list[SeriesResult]]:
|
|
||||||
global _CATALOG_INDEX_MEMORY
|
|
||||||
expires_at, cached = _CATALOG_INDEX_MEMORY
|
|
||||||
if cached and expires_at > time.time():
|
|
||||||
return list(cached)
|
|
||||||
raw = _session_cache_get(CATALOG_SEARCH_CACHE_KEY)
|
|
||||||
if not isinstance(raw, list):
|
|
||||||
return None
|
|
||||||
items: list[SeriesResult] = []
|
|
||||||
for entry in raw:
|
|
||||||
if not isinstance(entry, list) or len(entry) < 2:
|
|
||||||
continue
|
|
||||||
title = str(entry[0] or "").strip()
|
|
||||||
url = str(entry[1] or "").strip()
|
|
||||||
description = str(entry[2] or "") if len(entry) > 2 else ""
|
|
||||||
cover = str(entry[3] or "").strip() if len(entry) > 3 else ""
|
|
||||||
if title and url:
|
|
||||||
items.append(SeriesResult(title=title, description=description, url=url, cover=cover))
|
|
||||||
if items:
|
|
||||||
_CATALOG_INDEX_MEMORY = (time.time() + CATALOG_SEARCH_TTL_SECONDS, list(items))
|
|
||||||
return items or None
|
|
||||||
|
|
||||||
|
|
||||||
def _store_catalog_index_in_cache(items: list[SeriesResult]) -> None:
|
|
||||||
global _CATALOG_INDEX_MEMORY
|
|
||||||
if not items:
|
|
||||||
return
|
|
||||||
_CATALOG_INDEX_MEMORY = (time.time() + CATALOG_SEARCH_TTL_SECONDS, list(items))
|
|
||||||
payload: list[list[str]] = []
|
|
||||||
for entry in items:
|
|
||||||
if not entry.title or not entry.url:
|
|
||||||
continue
|
|
||||||
payload.append([entry.title, entry.url, entry.description, entry.cover])
|
|
||||||
_session_cache_set(CATALOG_SEARCH_CACHE_KEY, payload, ttl_seconds=CATALOG_SEARCH_TTL_SECONDS)
|
|
||||||
|
|
||||||
|
|
||||||
def search_series(query: str, *, progress_callback: ProgressCallback = None) -> list[SeriesResult]:
|
def search_series(query: str, *, progress_callback: ProgressCallback = None) -> list[SeriesResult]:
|
||||||
"""Sucht Serien. Katalog-Suche (vollstaendig) oder API-Suche (max 10) je nach Setting."""
|
"""Sucht Serien. Server-Suche (/suche?term=) zuerst, API als Fallback."""
|
||||||
_ensure_requests()
|
_ensure_requests()
|
||||||
if not _normalize_search_text(query):
|
if not _normalize_search_text(query):
|
||||||
return []
|
return []
|
||||||
|
|
||||||
use_catalog = _get_setting_bool(SETTING_CATALOG_SEARCH, default=True)
|
# 1. Server-Suche (schnell, vollstaendig, direkte HTML-Suche)
|
||||||
|
_emit_progress(progress_callback, "Suche", 20)
|
||||||
if use_catalog:
|
|
||||||
_emit_progress(progress_callback, "Pruefe Such-Cache", 15)
|
|
||||||
cached = _load_catalog_index_from_cache()
|
|
||||||
if cached is not None:
|
|
||||||
matched_from_cache = [entry for entry in cached if entry.title and _matches_query(query, title=entry.title)]
|
|
||||||
_emit_progress(progress_callback, f"Cache-Treffer: {len(cached)}", 35)
|
|
||||||
if matched_from_cache:
|
|
||||||
return matched_from_cache
|
|
||||||
|
|
||||||
_emit_progress(progress_callback, "Lade Katalogseite", 42)
|
|
||||||
catalog_url = f"{_get_base_url()}/serien?by=genre"
|
|
||||||
items: list[SeriesResult] = []
|
|
||||||
try:
|
|
||||||
soup = _get_soup_simple(catalog_url)
|
|
||||||
items = _catalog_index_from_soup(soup)
|
|
||||||
except Exception:
|
|
||||||
body = _get_html_simple(catalog_url)
|
|
||||||
items = _extract_catalog_index_from_html(body, progress_callback=progress_callback)
|
|
||||||
if not items:
|
|
||||||
_emit_progress(progress_callback, "Fallback-Parser", 58)
|
|
||||||
soup = BeautifulSoup(body, "html.parser")
|
|
||||||
items = _catalog_index_from_soup(soup)
|
|
||||||
if items:
|
|
||||||
_store_catalog_index_in_cache(items)
|
|
||||||
_emit_progress(progress_callback, f"Filtere Treffer ({len(items)})", 70)
|
|
||||||
return [entry for entry in items if entry.title and _matches_query(query, title=entry.title)]
|
|
||||||
|
|
||||||
# API-Suche (primaer wenn Katalog deaktiviert, Fallback wenn Katalog leer)
|
|
||||||
_emit_progress(progress_callback, "API-Suche", 60)
|
|
||||||
api_results = _search_series_api(query)
|
|
||||||
if api_results:
|
|
||||||
_emit_progress(progress_callback, f"API-Treffer: {len(api_results)}", 80)
|
|
||||||
return api_results
|
|
||||||
|
|
||||||
_emit_progress(progress_callback, "Server-Suche", 85)
|
|
||||||
server_results = _search_series_server(query)
|
server_results = _search_series_server(query)
|
||||||
if server_results:
|
if server_results:
|
||||||
_emit_progress(progress_callback, f"Server-Treffer: {len(server_results)}", 95)
|
return server_results
|
||||||
return [entry for entry in server_results if entry.title and _matches_query(query, title=entry.title)]
|
|
||||||
return []
|
# 2. API-Suche (Fallback, max 10 Ergebnisse)
|
||||||
|
_emit_progress(progress_callback, "API-Suche", 60)
|
||||||
|
return _search_series_api(query)
|
||||||
|
|
||||||
|
|
||||||
def parse_series_catalog(soup: BeautifulSoupT) -> dict[str, list[SeriesResult]]:
|
def parse_series_catalog(soup: BeautifulSoupT) -> dict[str, list[SeriesResult]]:
|
||||||
@@ -1159,6 +1027,7 @@ class SerienstreamPlugin(BasisPlugin):
|
|||||||
self._latest_hoster_cache: dict[str, list[str]] = {}
|
self._latest_hoster_cache: dict[str, list[str]] = {}
|
||||||
self._series_metadata_cache: dict[str, tuple[dict[str, str], dict[str, str]]] = {}
|
self._series_metadata_cache: dict[str, tuple[dict[str, str], dict[str, str]]] = {}
|
||||||
self._series_metadata_full: set[str] = set()
|
self._series_metadata_full: set[str] = set()
|
||||||
|
self._collection_url_cache: dict[str, str] = {}
|
||||||
self.is_available = True
|
self.is_available = True
|
||||||
self.unavailable_reason: str | None = None
|
self.unavailable_reason: str | None = None
|
||||||
if not self._requests_available: # pragma: no cover - optional dependency
|
if not self._requests_available: # pragma: no cover - optional dependency
|
||||||
@@ -1252,7 +1121,7 @@ class SerienstreamPlugin(BasisPlugin):
|
|||||||
except Exception:
|
except Exception:
|
||||||
continue
|
continue
|
||||||
url = str(item.get("url") or "").strip()
|
url = str(item.get("url") or "").strip()
|
||||||
if number <= 0 or not url:
|
if number < 0 or not url:
|
||||||
continue
|
continue
|
||||||
seasons.append(SeasonInfo(number=number, url=url, episodes=[]))
|
seasons.append(SeasonInfo(number=number, url=url, episodes=[]))
|
||||||
if not seasons:
|
if not seasons:
|
||||||
@@ -1383,7 +1252,60 @@ class SerienstreamPlugin(BasisPlugin):
|
|||||||
|
|
||||||
def capabilities(self) -> set[str]:
|
def capabilities(self) -> set[str]:
|
||||||
"""Meldet unterstützte Features für Router-Menüs."""
|
"""Meldet unterstützte Features für Router-Menüs."""
|
||||||
return {"popular_series", "genres", "latest_episodes", "alpha"}
|
return {"popular_series", "genres", "latest_episodes", "alpha", "collections"}
|
||||||
|
|
||||||
|
def collections(self) -> list[str]:
|
||||||
|
"""Liefert alle Sammlungs-Namen von /sammlungen (alle Seiten)."""
|
||||||
|
if not self._requests_available:
|
||||||
|
return []
|
||||||
|
base = _get_base_url()
|
||||||
|
names: list[str] = []
|
||||||
|
url_map: dict[str, str] = {}
|
||||||
|
page = 1
|
||||||
|
while True:
|
||||||
|
url = f"{base}/sammlungen" if page == 1 else f"{base}/sammlungen?page={page}"
|
||||||
|
soup = _get_soup_simple(url)
|
||||||
|
found = False
|
||||||
|
for a in soup.select('a[href*="/sammlung/"]'):
|
||||||
|
h2 = a.find("h2")
|
||||||
|
if not h2:
|
||||||
|
continue
|
||||||
|
title = h2.get_text(strip=True)
|
||||||
|
href = (a.get("href") or "").strip()
|
||||||
|
if title and href:
|
||||||
|
url_map[title] = _absolute_url(href)
|
||||||
|
names.append(title)
|
||||||
|
found = True
|
||||||
|
if not found:
|
||||||
|
break
|
||||||
|
if not soup.select(f'a[href*="/sammlungen?page={page + 1}"]'):
|
||||||
|
break
|
||||||
|
page += 1
|
||||||
|
if url_map:
|
||||||
|
_session_cache_set("collection_urls", url_map)
|
||||||
|
return names
|
||||||
|
|
||||||
|
def titles_for_collection(self, collection: str, page: int = 1) -> list[str]:
|
||||||
|
"""Liefert Serien-Titel einer Sammlung (paginiert)."""
|
||||||
|
if not self._requests_available:
|
||||||
|
return []
|
||||||
|
url_map = _session_cache_get("collection_urls")
|
||||||
|
if isinstance(url_map, dict):
|
||||||
|
self._collection_url_cache.update(url_map)
|
||||||
|
url = self._collection_url_cache.get(collection)
|
||||||
|
if not url:
|
||||||
|
return []
|
||||||
|
if page > 1:
|
||||||
|
url = f"{url}?page={page}"
|
||||||
|
soup = _get_soup_simple(url)
|
||||||
|
titles: list[str] = []
|
||||||
|
for a in soup.select('h6 a[href*="/serie/"]'):
|
||||||
|
title = a.get_text(strip=True)
|
||||||
|
href = (a.get("href") or "").strip()
|
||||||
|
if title and href:
|
||||||
|
self._remember_series_result(title, _absolute_url(href), "")
|
||||||
|
titles.append(title)
|
||||||
|
return titles
|
||||||
|
|
||||||
def popular_series(self) -> list[str]:
|
def popular_series(self) -> list[str]:
|
||||||
"""Liefert die Titel der beliebten Serien (Quelle: `/beliebte-serien`)."""
|
"""Liefert die Titel der beliebten Serien (Quelle: `/beliebte-serien`)."""
|
||||||
@@ -1794,6 +1716,8 @@ class SerienstreamPlugin(BasisPlugin):
|
|||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def _season_label(number: int) -> str:
|
def _season_label(number: int) -> str:
|
||||||
|
if number == 0:
|
||||||
|
return "Filme"
|
||||||
return f"Staffel {number}"
|
return f"Staffel {number}"
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
@@ -1808,6 +1732,8 @@ class SerienstreamPlugin(BasisPlugin):
|
|||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def _parse_season_number(label: str) -> int | None:
|
def _parse_season_number(label: str) -> int | None:
|
||||||
|
if (label or "").strip().casefold() == "filme":
|
||||||
|
return 0
|
||||||
digits = "".join(ch for ch in label if ch.isdigit())
|
digits = "".join(ch for ch in label if ch.isdigit())
|
||||||
if not digits:
|
if not digits:
|
||||||
return None
|
return None
|
||||||
|
|||||||
@@ -2,8 +2,7 @@
|
|||||||
<settings>
|
<settings>
|
||||||
<category label="Quellen">
|
<category label="Quellen">
|
||||||
<setting id="serienstream_base_url" type="text" label="SerienStream Basis-URL" default="https://s.to" />
|
<setting id="serienstream_base_url" type="text" label="SerienStream Basis-URL" default="https://s.to" />
|
||||||
<setting id="serienstream_catalog_search" type="bool" label="SerienStream: Katalog-Suche (mehr Ergebnisse, langsamer)" default="true" />
|
<setting id="aniworld_base_url" type="text" label="AniWorld Basis-URL" default="https://aniworld.to" />
|
||||||
<setting id="aniworld_base_url" type="text" label="AniWorld Basis-URL" default="https://aniworld.to" />
|
|
||||||
<setting id="topstream_base_url" type="text" label="TopStream Basis-URL" default="https://topstreamfilm.live" />
|
<setting id="topstream_base_url" type="text" label="TopStream Basis-URL" default="https://topstreamfilm.live" />
|
||||||
<setting id="einschalten_base_url" type="text" label="Einschalten Basis-URL" default="https://einschalten.in" />
|
<setting id="einschalten_base_url" type="text" label="Einschalten Basis-URL" default="https://einschalten.in" />
|
||||||
<setting id="filmpalast_base_url" type="text" label="Filmpalast Basis-URL" default="https://filmpalast.to" />
|
<setting id="filmpalast_base_url" type="text" label="Filmpalast Basis-URL" default="https://filmpalast.to" />
|
||||||
|
|||||||
Reference in New Issue
Block a user