Compare commits
2 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
| ae01c79555 | |||
| 22b9ae9c31 |
@@ -1,11 +0,0 @@
|
||||
# Changelog (Dev)
|
||||
|
||||
## 0.1.62-dev - 2026-02-24
|
||||
|
||||
- Neuer Dev-Stand fuer Genre-Performance (Serienstream).
|
||||
- Genre-Listen laden strikt nur die angeforderte Seite (on-demand, max. 20 Titel).
|
||||
- Weitere Seiten werden erst bei `Naechste Seite` geladen.
|
||||
- Listen-Parser reduziert auf Titel, Serien-URL und Cover.
|
||||
- Plot wird aus den Karten mit uebernommen und in der Liste angezeigt, falls vorhanden.
|
||||
- Metadaten werden fuer die jeweils geoeffnete Seite vollstaendig geladen und angezeigt.
|
||||
- Serien-Infos (inkl. Plot/Art) sind bereits in der Titelauswahl sichtbar, nicht erst in der Staffelansicht.
|
||||
@@ -1,14 +1,5 @@
|
||||
# Changelog (Nightly)
|
||||
|
||||
## 0.1.62-nightly - 2026-02-24
|
||||
|
||||
- Serienstream Genres auf strict on-demand Paging umgestellt:
|
||||
- Beim Oeffnen eines Genres wird nur Seite 1 geladen (max. 20 Titel).
|
||||
- Weitere Seiten werden nur bei `Naechste Seite` geladen.
|
||||
- Listen-Parser fuer Serienstream auf Titel, Serien-URL, Cover und Plot optimiert.
|
||||
- Serien-Infos (Plot/Art) sind bereits in der Titelauswahl sichtbar.
|
||||
- Dev-Changelog-Datei eingefuehrt (`CHANGELOG-DEV.md`) fuer `-dev` Builds.
|
||||
|
||||
## 0.1.61-nightly - 2026-02-23
|
||||
|
||||
- Update-Dialog: feste Auswahl mit `Installieren` / `Abbrechen` (kein vertauschter Yes/No-Dialog mehr).
|
||||
|
||||
11
CHANGELOG.md
11
CHANGELOG.md
@@ -1,5 +1,16 @@
|
||||
# Changelog (Stable)
|
||||
|
||||
## 0.1.61 - 2026-02-23
|
||||
|
||||
- Menues und Labels weiter vereinheitlicht (ASCII-only, einheitliche Texte pro Plugin).
|
||||
- Update-Bereich ueberarbeitet:
|
||||
- Kanalwechsel mit direkter Installation der neuesten Kanal-Version.
|
||||
- Version-Auswahl mit Changelog-Anzeige und klarer Installieren/Abbrechen-Auswahl.
|
||||
- Anzeige der installierten Version direkt aus lokaler `addon.xml`.
|
||||
- Kanal-spezifischer Versionsfilter (Main nur stable, Nightly nur `-nightly`).
|
||||
- Resolver-/Playback-Flow vereinheitlicht und Hoster-URL-Normalisierung zentralisiert.
|
||||
- Settings aufgeraeumt (strukturierte Kategorien, reduzierte Alt-Optionen).
|
||||
|
||||
## 0.1.58 - 2026-02-23
|
||||
|
||||
- Menuebezeichnungen vereinheitlicht (`Haeufig gesehen`, `Neuste Titel`).
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
<?xml version='1.0' encoding='utf-8'?>
|
||||
<addon id="plugin.video.viewit" name="ViewIt" version="0.1.62-nightly" provider-name="ViewIt">
|
||||
<addon id="plugin.video.viewit" name="ViewIt" version="0.1.61" provider-name="ViewIt">
|
||||
<requires>
|
||||
<import addon="xbmc.python" version="3.0.0" />
|
||||
<import addon="script.module.requests" />
|
||||
|
||||
@@ -1170,12 +1170,7 @@ def _extract_changelog_section(changelog_text: str, version: str) -> str:
|
||||
|
||||
|
||||
def _fetch_changelog_for_channel(channel: int, version: str) -> str:
|
||||
version_text = str(version or "").strip().casefold()
|
||||
if version_text.endswith("-dev"):
|
||||
url = "https://gitea.it-drui.de/viewit/ViewIT/raw/branch/dev/CHANGELOG-DEV.md"
|
||||
elif version_text.endswith("-nightly"):
|
||||
url = "https://gitea.it-drui.de/viewit/ViewIT/raw/branch/nightly/CHANGELOG-NIGHTLY.md"
|
||||
elif channel == UPDATE_CHANNEL_MAIN:
|
||||
if channel == UPDATE_CHANNEL_MAIN:
|
||||
url = "https://gitea.it-drui.de/viewit/ViewIT/raw/branch/main/CHANGELOG.md"
|
||||
else:
|
||||
url = "https://gitea.it-drui.de/viewit/ViewIT/raw/branch/nightly/CHANGELOG-NIGHTLY.md"
|
||||
|
||||
@@ -79,7 +79,6 @@ SESSION_CACHE_PREFIX = "viewit.serienstream"
|
||||
SESSION_CACHE_MAX_TITLE_URLS = 800
|
||||
CATALOG_SEARCH_TTL_SECONDS = 600
|
||||
CATALOG_SEARCH_CACHE_KEY = "catalog_index"
|
||||
GENRE_LIST_PAGE_SIZE = 20
|
||||
_CATALOG_INDEX_MEMORY: tuple[float, List["SeriesResult"]] = (0.0, [])
|
||||
ProgressCallback = Optional[Callable[[str, Optional[int]], Any]]
|
||||
|
||||
@@ -98,7 +97,6 @@ class SeriesResult:
|
||||
title: str
|
||||
description: str
|
||||
url: str
|
||||
cover: str = ""
|
||||
|
||||
|
||||
@dataclass
|
||||
@@ -671,9 +669,8 @@ def _load_catalog_index_from_cache() -> Optional[List[SeriesResult]]:
|
||||
title = str(entry[0] or "").strip()
|
||||
url = str(entry[1] or "").strip()
|
||||
description = str(entry[2] or "") if len(entry) > 2 else ""
|
||||
cover = str(entry[3] or "").strip() if len(entry) > 3 else ""
|
||||
if title and url:
|
||||
items.append(SeriesResult(title=title, description=description, url=url, cover=cover))
|
||||
items.append(SeriesResult(title=title, description=description, url=url))
|
||||
if items:
|
||||
_CATALOG_INDEX_MEMORY = (time.time() + CATALOG_SEARCH_TTL_SECONDS, list(items))
|
||||
return items or None
|
||||
@@ -688,7 +685,7 @@ def _store_catalog_index_in_cache(items: List[SeriesResult]) -> None:
|
||||
for entry in items:
|
||||
if not entry.title or not entry.url:
|
||||
continue
|
||||
payload.append([entry.title, entry.url, entry.description, entry.cover])
|
||||
payload.append([entry.title, entry.url, entry.description])
|
||||
_session_cache_set(CATALOG_SEARCH_CACHE_KEY, payload, ttl_seconds=CATALOG_SEARCH_TTL_SECONDS)
|
||||
|
||||
|
||||
@@ -1110,8 +1107,8 @@ class SerienstreamPlugin(BasisPlugin):
|
||||
self._episode_label_cache: Dict[Tuple[str, str], Dict[str, EpisodeInfo]] = {}
|
||||
self._catalog_cache: Optional[Dict[str, List[SeriesResult]]] = None
|
||||
self._genre_group_cache: Dict[str, Dict[str, List[str]]] = {}
|
||||
self._genre_page_entries_cache: Dict[Tuple[str, int], List[SeriesResult]] = {}
|
||||
self._genre_page_has_more_cache: Dict[Tuple[str, int], bool] = {}
|
||||
self._genre_page_titles_cache: Dict[Tuple[str, int], List[str]] = {}
|
||||
self._genre_page_count_cache: Dict[str, int] = {}
|
||||
self._popular_cache: Optional[List[SeriesResult]] = None
|
||||
self._requests_available = REQUESTS_AVAILABLE
|
||||
self._default_preferred_hosters: List[str] = list(DEFAULT_PREFERRED_HOSTERS)
|
||||
@@ -1120,7 +1117,6 @@ class SerienstreamPlugin(BasisPlugin):
|
||||
self._latest_cache: Dict[int, List[LatestEpisode]] = {}
|
||||
self._latest_hoster_cache: Dict[str, List[str]] = {}
|
||||
self._series_metadata_cache: Dict[str, Tuple[Dict[str, str], Dict[str, str]]] = {}
|
||||
self._series_metadata_full: set[str] = set()
|
||||
self.is_available = True
|
||||
self.unavailable_reason: Optional[str] = None
|
||||
if not self._requests_available: # pragma: no cover - optional dependency
|
||||
@@ -1413,165 +1409,49 @@ class SerienstreamPlugin(BasisPlugin):
|
||||
value = re.sub(r"[^a-z0-9]+", "-", value).strip("-")
|
||||
return value
|
||||
|
||||
def _cache_list_metadata(self, title: str, description: str = "", cover: str = "") -> None:
|
||||
key = self._metadata_cache_key(title)
|
||||
cached = self._series_metadata_cache.get(key)
|
||||
info = dict(cached[0]) if cached else {}
|
||||
art = dict(cached[1]) if cached else {}
|
||||
info.setdefault("title", title)
|
||||
description = (description or "").strip()
|
||||
if description and not info.get("plot"):
|
||||
info["plot"] = description
|
||||
cover = _absolute_url((cover or "").strip()) if cover else ""
|
||||
if cover:
|
||||
art.setdefault("thumb", cover)
|
||||
art.setdefault("poster", cover)
|
||||
self._series_metadata_cache[key] = (info, art)
|
||||
|
||||
@staticmethod
|
||||
def _card_description(anchor: BeautifulSoupT) -> str:
|
||||
if not anchor:
|
||||
return ""
|
||||
candidates: List[str] = []
|
||||
direct = (anchor.get("data-search") or "").strip()
|
||||
if direct:
|
||||
candidates.append(direct)
|
||||
title_attr = (anchor.get("data-title") or "").strip()
|
||||
if title_attr:
|
||||
candidates.append(title_attr)
|
||||
for selector in ("p", ".description", ".desc", ".text-muted", ".small", ".overview"):
|
||||
node = anchor.select_one(selector)
|
||||
if node is None:
|
||||
continue
|
||||
text = (node.get_text(" ", strip=True) or "").strip()
|
||||
if text:
|
||||
candidates.append(text)
|
||||
parent = anchor.parent if anchor else None
|
||||
if parent is not None:
|
||||
parent_data = (parent.get("data-search") or "").strip()
|
||||
if parent_data:
|
||||
candidates.append(parent_data)
|
||||
parent_text = ""
|
||||
try:
|
||||
parent_text = (parent.get_text(" ", strip=True) or "").strip()
|
||||
except Exception:
|
||||
parent_text = ""
|
||||
if parent_text and len(parent_text) > 24:
|
||||
candidates.append(parent_text)
|
||||
for value in candidates:
|
||||
cleaned = re.sub(r"\s+", " ", str(value or "")).strip()
|
||||
if cleaned and len(cleaned) > 12:
|
||||
return cleaned
|
||||
return ""
|
||||
|
||||
def _parse_genre_entries_from_soup(self, soup: BeautifulSoupT) -> List[SeriesResult]:
|
||||
entries: List[SeriesResult] = []
|
||||
seen_urls: set[str] = set()
|
||||
|
||||
def _add_entry(title: str, description: str, href: str, cover: str) -> None:
|
||||
series_url = _absolute_url(href).split("#", 1)[0].split("?", 1)[0].rstrip("/")
|
||||
if not series_url or "/serie/" not in series_url:
|
||||
return
|
||||
if "/staffel-" in series_url or "/episode-" in series_url:
|
||||
return
|
||||
if series_url in seen_urls:
|
||||
return
|
||||
title = (title or "").strip()
|
||||
if not title:
|
||||
return
|
||||
description = (description or "").strip()
|
||||
cover_url = _absolute_url((cover or "").strip()) if cover else ""
|
||||
seen_urls.add(series_url)
|
||||
self._remember_series_result(title, series_url, description)
|
||||
self._cache_list_metadata(title, description=description, cover=cover_url)
|
||||
entries.append(SeriesResult(title=title, description=description, url=series_url, cover=cover_url))
|
||||
|
||||
for anchor in soup.select("a.show-card[href]"):
|
||||
href = (anchor.get("href") or "").strip()
|
||||
if not href:
|
||||
continue
|
||||
img = anchor.select_one("img")
|
||||
title = (
|
||||
(img.get("alt") if img else "")
|
||||
or (anchor.get("title") or "")
|
||||
or (anchor.get_text(" ", strip=True) or "")
|
||||
).strip()
|
||||
description = self._card_description(anchor)
|
||||
cover = (img.get("data-src") if img else "") or (img.get("src") if img else "")
|
||||
_add_entry(title, description, href, cover)
|
||||
|
||||
if entries:
|
||||
return entries
|
||||
|
||||
for item in soup.select("li.series-item"):
|
||||
anchor = item.find("a", href=True)
|
||||
if not anchor:
|
||||
continue
|
||||
href = (anchor.get("href") or "").strip()
|
||||
title = (anchor.get_text(" ", strip=True) or "").strip()
|
||||
description = (item.get("data-search") or "").strip()
|
||||
img = anchor.find("img")
|
||||
cover = (img.get("data-src") if img else "") or (img.get("src") if img else "")
|
||||
_add_entry(title, description, href, cover)
|
||||
return entries
|
||||
|
||||
def _fetch_genre_page_entries(self, genre: str, page: int) -> Tuple[List[SeriesResult], bool]:
|
||||
def _fetch_genre_page_titles(self, genre: str, page: int) -> Tuple[List[str], int]:
|
||||
slug = self._genre_slug(genre)
|
||||
if not slug:
|
||||
return [], False
|
||||
return [], 1
|
||||
cache_key = (slug, page)
|
||||
cached_entries = self._genre_page_entries_cache.get(cache_key)
|
||||
cached_has_more = self._genre_page_has_more_cache.get(cache_key)
|
||||
if cached_entries is not None and cached_has_more is not None:
|
||||
return list(cached_entries), bool(cached_has_more)
|
||||
cached = self._genre_page_titles_cache.get(cache_key)
|
||||
cached_pages = self._genre_page_count_cache.get(slug)
|
||||
if cached is not None and cached_pages is not None:
|
||||
return list(cached), int(cached_pages)
|
||||
url = f"{_get_base_url()}/genre/{slug}"
|
||||
if page > 1:
|
||||
url = f"{url}?page={int(page)}"
|
||||
soup = _get_soup_simple(url)
|
||||
entries = self._parse_genre_entries_from_soup(soup)
|
||||
|
||||
has_more = False
|
||||
for anchor in soup.select("a[rel='next'][href], a[href*='?page=']"):
|
||||
titles: List[str] = []
|
||||
seen: set[str] = set()
|
||||
for anchor in soup.select("a.show-card[href]"):
|
||||
href = (anchor.get("href") or "").strip()
|
||||
if not href:
|
||||
series_url = _absolute_url(href).split("#", 1)[0].split("?", 1)[0].rstrip("/")
|
||||
if "/serie/" not in series_url:
|
||||
continue
|
||||
img = anchor.select_one("img[alt]")
|
||||
title = ((img.get("alt") if img else "") or "").strip()
|
||||
if not title:
|
||||
continue
|
||||
key = title.casefold()
|
||||
if key in seen:
|
||||
continue
|
||||
seen.add(key)
|
||||
self._remember_series_result(title, series_url)
|
||||
titles.append(title)
|
||||
max_page = 1
|
||||
for anchor in soup.select("a[href*='?page=']"):
|
||||
href = (anchor.get("href") or "").strip()
|
||||
match = re.search(r"[?&]page=(\d+)", href)
|
||||
if not match:
|
||||
if "next" in href.casefold():
|
||||
has_more = True
|
||||
continue
|
||||
try:
|
||||
if int(match.group(1)) > int(page):
|
||||
has_more = True
|
||||
break
|
||||
max_page = max(max_page, int(match.group(1)))
|
||||
except Exception:
|
||||
continue
|
||||
if len(entries) > GENRE_LIST_PAGE_SIZE:
|
||||
has_more = True
|
||||
entries = entries[:GENRE_LIST_PAGE_SIZE]
|
||||
|
||||
self._genre_page_entries_cache[cache_key] = list(entries)
|
||||
self._genre_page_has_more_cache[cache_key] = bool(has_more)
|
||||
return list(entries), bool(has_more)
|
||||
|
||||
def titles_for_genre_page(self, genre: str, page: int) -> List[str]:
|
||||
genre = (genre or "").strip()
|
||||
page = max(1, int(page or 1))
|
||||
entries, _ = self._fetch_genre_page_entries(genre, page)
|
||||
return [entry.title for entry in entries if entry.title]
|
||||
|
||||
def genre_has_more(self, genre: str, page: int) -> bool:
|
||||
genre = (genre or "").strip()
|
||||
page = max(1, int(page or 1))
|
||||
slug = self._genre_slug(genre)
|
||||
if not slug:
|
||||
return False
|
||||
cache_key = (slug, page)
|
||||
cached = self._genre_page_has_more_cache.get(cache_key)
|
||||
if cached is not None:
|
||||
return bool(cached)
|
||||
_, has_more = self._fetch_genre_page_entries(genre, page)
|
||||
return bool(has_more)
|
||||
self._genre_page_titles_cache[cache_key] = list(titles)
|
||||
self._genre_page_count_cache[slug] = max_page
|
||||
return list(titles), max_page
|
||||
|
||||
def titles_for_genre_group_page(self, genre: str, group_code: str, page: int = 1, page_size: int = 10) -> List[str]:
|
||||
genre = (genre or "").strip()
|
||||
@@ -1581,17 +1461,14 @@ class SerienstreamPlugin(BasisPlugin):
|
||||
needed = page * page_size + 1
|
||||
matched: List[str] = []
|
||||
try:
|
||||
page_index = 1
|
||||
has_more = True
|
||||
while has_more:
|
||||
page_entries, has_more = self._fetch_genre_page_entries(genre, page_index)
|
||||
for entry in page_entries:
|
||||
title = entry.title
|
||||
_, max_pages = self._fetch_genre_page_titles(genre, 1)
|
||||
for page_index in range(1, max_pages + 1):
|
||||
page_titles, _ = self._fetch_genre_page_titles(genre, page_index)
|
||||
for title in page_titles:
|
||||
if self._group_matches(group_code, title):
|
||||
matched.append(title)
|
||||
if len(matched) >= needed:
|
||||
break
|
||||
page_index += 1
|
||||
start = (page - 1) * page_size
|
||||
end = start + page_size
|
||||
return list(matched[start:end])
|
||||
@@ -1610,17 +1487,14 @@ class SerienstreamPlugin(BasisPlugin):
|
||||
needed = page * page_size + 1
|
||||
count = 0
|
||||
try:
|
||||
page_index = 1
|
||||
has_more = True
|
||||
while has_more:
|
||||
page_entries, has_more = self._fetch_genre_page_entries(genre, page_index)
|
||||
for entry in page_entries:
|
||||
title = entry.title
|
||||
_, max_pages = self._fetch_genre_page_titles(genre, 1)
|
||||
for page_index in range(1, max_pages + 1):
|
||||
page_titles, _ = self._fetch_genre_page_titles(genre, page_index)
|
||||
for title in page_titles:
|
||||
if self._group_matches(group_code, title):
|
||||
count += 1
|
||||
if count >= needed:
|
||||
return True
|
||||
page_index += 1
|
||||
return False
|
||||
except Exception:
|
||||
grouped = self._ensure_genre_group_cache(genre)
|
||||
@@ -1737,7 +1611,6 @@ class SerienstreamPlugin(BasisPlugin):
|
||||
cache_key = self._metadata_cache_key(title)
|
||||
if info_labels or art:
|
||||
self._series_metadata_cache[cache_key] = (info_labels, art)
|
||||
self._series_metadata_full.add(cache_key)
|
||||
|
||||
base_series_url = _series_root_url(_extract_canonical_url(series_soup, series.url))
|
||||
season_links = _extract_season_links(series_soup)
|
||||
@@ -1773,7 +1646,7 @@ class SerienstreamPlugin(BasisPlugin):
|
||||
|
||||
cache_key = self._metadata_cache_key(title)
|
||||
cached = self._series_metadata_cache.get(cache_key)
|
||||
if cached is not None and cache_key in self._series_metadata_full:
|
||||
if cached is not None:
|
||||
info, art = cached
|
||||
return dict(info), dict(art), None
|
||||
|
||||
@@ -1783,14 +1656,11 @@ class SerienstreamPlugin(BasisPlugin):
|
||||
self._series_metadata_cache[cache_key] = (dict(info), {})
|
||||
return info, {}, None
|
||||
|
||||
info: Dict[str, str] = dict(cached[0]) if cached else {"title": title}
|
||||
art: Dict[str, str] = dict(cached[1]) if cached else {}
|
||||
info.setdefault("title", title)
|
||||
info: Dict[str, str] = {"title": title}
|
||||
art: Dict[str, str] = {}
|
||||
if series.description:
|
||||
info.setdefault("plot", series.description)
|
||||
info["plot"] = series.description
|
||||
|
||||
# Fuer Listenansichten laden wir pro Seite die Detail-Metadaten vollstaendig nach.
|
||||
loaded_full = False
|
||||
try:
|
||||
soup = _get_soup(series.url, session=get_requests_session("serienstream", headers=HEADERS))
|
||||
parsed_info, parsed_art = _extract_series_metadata(soup)
|
||||
@@ -1798,13 +1668,10 @@ class SerienstreamPlugin(BasisPlugin):
|
||||
info.update(parsed_info)
|
||||
if parsed_art:
|
||||
art.update(parsed_art)
|
||||
loaded_full = True
|
||||
except Exception:
|
||||
pass
|
||||
|
||||
self._series_metadata_cache[cache_key] = (dict(info), dict(art))
|
||||
if loaded_full:
|
||||
self._series_metadata_full.add(cache_key)
|
||||
return info, art, None
|
||||
|
||||
def series_url_for_title(self, title: str) -> str:
|
||||
@@ -1875,8 +1742,6 @@ class SerienstreamPlugin(BasisPlugin):
|
||||
self._season_links_cache.clear()
|
||||
self._episode_label_cache.clear()
|
||||
self._catalog_cache = None
|
||||
self._series_metadata_cache.clear()
|
||||
self._series_metadata_full.clear()
|
||||
return []
|
||||
if not self._requests_available:
|
||||
raise RuntimeError("SerienstreamPlugin kann ohne requests/bs4 nicht suchen.")
|
||||
@@ -1890,8 +1755,6 @@ class SerienstreamPlugin(BasisPlugin):
|
||||
self._season_cache.clear()
|
||||
self._episode_label_cache.clear()
|
||||
self._catalog_cache = None
|
||||
self._series_metadata_cache.clear()
|
||||
self._series_metadata_full.clear()
|
||||
raise RuntimeError(f"Serienstream-Suche fehlgeschlagen: {exc}") from exc
|
||||
self._series_results = {}
|
||||
for result in results:
|
||||
|
||||
@@ -36,7 +36,7 @@
|
||||
</category>
|
||||
|
||||
<category label="Updates">
|
||||
<setting id="update_channel" type="enum" label="Update-Kanal" default="1" values="Main|Nightly|Custom" />
|
||||
<setting id="update_channel" type="enum" label="Update-Kanal" default="0" values="Main|Nightly|Custom" />
|
||||
<setting id="apply_update_channel" type="action" label="Update-Kanal jetzt anwenden" action="RunPlugin(plugin://plugin.video.viewit/?action=apply_update_channel)" option="close" />
|
||||
<setting id="auto_update_enabled" type="bool" label="Automatische Updates (beim Start pruefen)" default="false" />
|
||||
<setting id="select_update_version" type="action" label="Version waehlen und installieren" action="RunPlugin(plugin://plugin.video.viewit/?action=select_update_version)" option="close" />
|
||||
@@ -49,7 +49,7 @@
|
||||
<setting id="update_info" type="text" label="Updates laufen ueber den normalen Kodi-Update-Mechanismus." default="" enable="false" />
|
||||
<setting id="update_repo_url_main" type="text" label="Main URL (addons.xml)" default="https://gitea.it-drui.de/viewit/ViewIT-Kodi-Repo/raw/branch/main/addons.xml" />
|
||||
<setting id="update_repo_url_nightly" type="text" label="Nightly URL (addons.xml)" default="https://gitea.it-drui.de/viewit/ViewIT-Kodi-Repo/raw/branch/nightly/addons.xml" />
|
||||
<setting id="update_repo_url" type="text" label="Custom URL (addons.xml)" default="https://gitea.it-drui.de/viewit/ViewIT-Kodi-Repo/raw/branch/nightly/addons.xml" />
|
||||
<setting id="update_repo_url" type="text" label="Custom URL (addons.xml)" default="https://gitea.it-drui.de/viewit/ViewIT-Kodi-Repo/raw/branch/main/addons.xml" />
|
||||
<setting id="auto_update_last_ts" type="text" label="Auto-Update letzte Pruefung (intern)" default="0" visible="false" />
|
||||
<setting id="update_version_addon" type="text" label="ViewIT Version" default="-" visible="false" />
|
||||
<setting id="update_version_serienstream" type="text" label="SerienStream Version" default="-" visible="false" />
|
||||
|
||||
@@ -1,21 +1,17 @@
|
||||
# Release Flow (Main + Nightly + Dev)
|
||||
# Release Flow (Main + Nightly)
|
||||
|
||||
This project uses three release channels:
|
||||
This project uses two release channels:
|
||||
|
||||
- `dev`: playground for experiments
|
||||
- `nightly`: integration and test channel
|
||||
- `main`: stable channel
|
||||
|
||||
## Rules
|
||||
|
||||
- Experimental work goes to `dev`.
|
||||
- Feature work for release goes to `nightly`.
|
||||
- Feature work goes to `nightly` only.
|
||||
- Promote from `nightly` to `main` with `--squash` only.
|
||||
- `main` version has no suffix (`0.1.60`).
|
||||
- `nightly` version uses `-nightly` and is always at least one patch higher than `main` (`0.1.61-nightly`).
|
||||
- `dev` version uses `-dev` (`0.1.62-dev`).
|
||||
- Keep changelogs split:
|
||||
- `CHANGELOG-DEV.md`
|
||||
- `CHANGELOG-NIGHTLY.md`
|
||||
- `CHANGELOG.md`
|
||||
|
||||
@@ -44,6 +40,5 @@ Then:
|
||||
|
||||
## Local ZIPs (separated)
|
||||
|
||||
- Dev ZIP output: `dist/local_zips/dev/`
|
||||
- Main ZIP output: `dist/local_zips/main/`
|
||||
- Nightly ZIP output: `dist/local_zips/nightly/`
|
||||
|
||||
Reference in New Issue
Block a user