mirror of
https://github.com/unshackle-dl/unshackle.git
synced 2025-10-23 15:11:08 +00:00
Compare commits
20 Commits
3ef43afeed
...
1.4.5
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
6137146705 | ||
|
|
859d09693c | ||
|
|
5f022635cb | ||
|
|
ad66502c0c | ||
|
|
e462f07b7a | ||
|
|
83b600e999 | ||
|
|
ea8a7b00c9 | ||
|
|
16ee4175a4 | ||
|
|
f722ec69b6 | ||
|
|
2330297ea4 | ||
|
|
86bb162868 | ||
|
|
501cfd68e8 | ||
|
|
76fb2eea95 | ||
|
|
ea5ec40bcd | ||
|
|
329850b043 | ||
|
|
73595f3b50 | ||
|
|
1e82283133 | ||
|
|
ab13dde9d2 | ||
|
|
9fd0895128 | ||
|
|
ed744205ad |
68
CHANGELOG.md
68
CHANGELOG.md
@@ -5,6 +5,74 @@ All notable changes to this project will be documented in this file.
|
|||||||
The format is based on [Keep a Changelog](https://keepachangelog.com/en/1.1.0/),
|
The format is based on [Keep a Changelog](https://keepachangelog.com/en/1.1.0/),
|
||||||
and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0.html).
|
and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0.html).
|
||||||
|
|
||||||
|
## [1.4.5] - 2025-09-09
|
||||||
|
|
||||||
|
### Added
|
||||||
|
|
||||||
|
- **Enhanced CDM Key Caching**: Improved key caching and session management for L1/L2 devices
|
||||||
|
- Optimized `get_cached_keys_if_exists` functionality for better performance with L1/L2 devices
|
||||||
|
- Enhanced cached key retrieval logic with improved session handling
|
||||||
|
- **Widevine Common Certificate Fallback**: Added fallback to Widevine common certificate for L1 devices
|
||||||
|
- Improved compatibility for L1 devices when service certificates are unavailable
|
||||||
|
- **Enhanced Vault Loading**: Improved vault loading and key copying logic
|
||||||
|
- Better error handling and key management in vault operations
|
||||||
|
- **PSSH Display Optimization**: Truncated PSSH string display in non-debug mode for cleaner output
|
||||||
|
- **CDM Error Messaging**: Added error messages for missing service certificates in CDM sessions
|
||||||
|
|
||||||
|
### Changed
|
||||||
|
|
||||||
|
- **Dynamic Version Headers**: Updated User-Agent headers to use dynamic version strings
|
||||||
|
- DecryptLabsRemoteCDM now uses dynamic version import instead of hardcoded version
|
||||||
|
- **Intelligent CDM Caching**: Implemented intelligent caching system for CDM license requests
|
||||||
|
- Enhanced caching logic reduces redundant license requests and improves performance
|
||||||
|
- **Enhanced Tag Handling**: Improved tag handling for TV shows and movies from Simkl data
|
||||||
|
- Better metadata processing and formatting for improved media tagging
|
||||||
|
|
||||||
|
### Fixed
|
||||||
|
|
||||||
|
- **CDM Session Management**: Clean up session data when retrieving cached keys
|
||||||
|
- Remove decrypt_labs_session_id and challenge from session when cached keys exist but there are missing kids
|
||||||
|
- Ensures clean state for subsequent requests and prevents session conflicts
|
||||||
|
- **Tag Formatting**: Fixed formatting issues in tag processing
|
||||||
|
- **Import Order**: Fixed import order issues in tags module
|
||||||
|
|
||||||
|
## [1.4.4] - 2025-09-02
|
||||||
|
|
||||||
|
### Added
|
||||||
|
|
||||||
|
- **Enhanced DecryptLabs CDM Support**: Comprehensive remote CDM functionality
|
||||||
|
- Full support for Widevine, PlayReady, and ChromeCDM through DecryptLabsRemoteCDM
|
||||||
|
- Enhanced session management and caching support for remote WV/PR operations
|
||||||
|
- Support for cached keys and improved license handling
|
||||||
|
- New CDM configurations for Chrome and PlayReady devices with updated User-Agent and service certificate
|
||||||
|
- **Advanced Configuration Options**: New device and language preferences
|
||||||
|
- Added configuration options for device certificate status list
|
||||||
|
- Enhanced language preference settings
|
||||||
|
|
||||||
|
### Changed
|
||||||
|
|
||||||
|
- **DRM Decryption Enhancements**: Streamlined decryption process
|
||||||
|
- Simplified decrypt method by removing unused parameter and streamlined logic
|
||||||
|
- Improved DecryptLabs CDM configurations with better device support
|
||||||
|
|
||||||
|
### Fixed
|
||||||
|
|
||||||
|
- **Matroska Tag Compliance**: Enhanced media container compatibility
|
||||||
|
- Fixed Matroska tag compliance with official specification
|
||||||
|
- **Application Branding**: Cleaned up version display
|
||||||
|
- Removed old devine version reference from banner to avoid developer confusion
|
||||||
|
- Updated branding while maintaining original GNU license compliance
|
||||||
|
- **IP Information Handling**: Improved geolocation services
|
||||||
|
- Enhanced get_ip_info functionality with better failover handling
|
||||||
|
- Added support for 429 error handling and multiple API provider fallback
|
||||||
|
- Implemented cached IP info retrieval with fallback tester to avoid rate limiting
|
||||||
|
- **Dependencies**: Streamlined package requirements
|
||||||
|
- Removed unnecessary data extra requirement from langcodes
|
||||||
|
|
||||||
|
### Removed
|
||||||
|
|
||||||
|
- Deprecated version references in application banner for clarity
|
||||||
|
|
||||||
## [1.4.3] - 2025-08-20
|
## [1.4.3] - 2025-08-20
|
||||||
|
|
||||||
### Added
|
### Added
|
||||||
|
|||||||
@@ -4,7 +4,7 @@ build-backend = "hatchling.build"
|
|||||||
|
|
||||||
[project]
|
[project]
|
||||||
name = "unshackle"
|
name = "unshackle"
|
||||||
version = "1.4.3"
|
version = "1.4.4"
|
||||||
description = "Modular Movie, TV, and Music Archival Software."
|
description = "Modular Movie, TV, and Music Archival Software."
|
||||||
authors = [{ name = "unshackle team" }]
|
authors = [{ name = "unshackle team" }]
|
||||||
requires-python = ">=3.10,<3.13"
|
requires-python = ">=3.10,<3.13"
|
||||||
|
|||||||
@@ -66,6 +66,18 @@ from unshackle.core.vaults import Vaults
|
|||||||
|
|
||||||
|
|
||||||
class dl:
|
class dl:
|
||||||
|
@staticmethod
|
||||||
|
def _truncate_pssh_for_display(pssh_string: str, drm_type: str) -> str:
|
||||||
|
"""Truncate PSSH string for display when not in debug mode."""
|
||||||
|
if logging.root.level == logging.DEBUG or not pssh_string:
|
||||||
|
return pssh_string
|
||||||
|
|
||||||
|
max_width = console.width - len(drm_type) - 12
|
||||||
|
if len(pssh_string) <= max_width:
|
||||||
|
return pssh_string
|
||||||
|
|
||||||
|
return pssh_string[: max_width - 3] + "..."
|
||||||
|
|
||||||
@click.command(
|
@click.command(
|
||||||
short_help="Download, Decrypt, and Mux tracks for titles from a Service.",
|
short_help="Download, Decrypt, and Mux tracks for titles from a Service.",
|
||||||
cls=Services,
|
cls=Services,
|
||||||
@@ -345,7 +357,10 @@ class dl:
|
|||||||
sys.exit(1)
|
sys.exit(1)
|
||||||
|
|
||||||
if self.cdm:
|
if self.cdm:
|
||||||
if hasattr(self.cdm, "device_type") and self.cdm.device_type.name in ["ANDROID", "CHROME"]:
|
if isinstance(self.cdm, DecryptLabsRemoteCDM):
|
||||||
|
drm_type = "PlayReady" if self.cdm.is_playready else "Widevine"
|
||||||
|
self.log.info(f"Loaded {drm_type} Remote CDM: DecryptLabs (L{self.cdm.security_level})")
|
||||||
|
elif hasattr(self.cdm, "device_type") and self.cdm.device_type.name in ["ANDROID", "CHROME"]:
|
||||||
self.log.info(f"Loaded Widevine CDM: {self.cdm.system_id} (L{self.cdm.security_level})")
|
self.log.info(f"Loaded Widevine CDM: {self.cdm.system_id} (L{self.cdm.security_level})")
|
||||||
else:
|
else:
|
||||||
self.log.info(
|
self.log.info(
|
||||||
@@ -874,7 +889,12 @@ class dl:
|
|||||||
),
|
),
|
||||||
licence=partial(
|
licence=partial(
|
||||||
service.get_playready_license
|
service.get_playready_license
|
||||||
if isinstance(self.cdm, PlayReadyCdm)
|
if (
|
||||||
|
isinstance(self.cdm, PlayReadyCdm)
|
||||||
|
or (
|
||||||
|
isinstance(self.cdm, DecryptLabsRemoteCDM) and self.cdm.is_playready
|
||||||
|
)
|
||||||
|
)
|
||||||
and hasattr(service, "get_playready_license")
|
and hasattr(service, "get_playready_license")
|
||||||
else service.get_widevine_license,
|
else service.get_widevine_license,
|
||||||
title=title,
|
title=title,
|
||||||
@@ -1201,14 +1221,27 @@ class dl:
|
|||||||
if not drm:
|
if not drm:
|
||||||
return
|
return
|
||||||
|
|
||||||
if isinstance(drm, Widevine) and not isinstance(self.cdm, WidevineCdm):
|
if isinstance(drm, Widevine):
|
||||||
self.cdm = self.get_cdm(self.service, self.profile, drm="widevine")
|
if not isinstance(self.cdm, (WidevineCdm, DecryptLabsRemoteCDM)) or (
|
||||||
elif isinstance(drm, PlayReady) and not isinstance(self.cdm, PlayReadyCdm):
|
isinstance(self.cdm, DecryptLabsRemoteCDM) and self.cdm.is_playready
|
||||||
self.cdm = self.get_cdm(self.service, self.profile, drm="playready")
|
):
|
||||||
|
widevine_cdm = self.get_cdm(self.service, self.profile, drm="widevine")
|
||||||
|
if widevine_cdm:
|
||||||
|
self.log.info("Switching to Widevine CDM for Widevine content")
|
||||||
|
self.cdm = widevine_cdm
|
||||||
|
elif isinstance(drm, PlayReady):
|
||||||
|
if not isinstance(self.cdm, (PlayReadyCdm, DecryptLabsRemoteCDM)) or (
|
||||||
|
isinstance(self.cdm, DecryptLabsRemoteCDM) and not self.cdm.is_playready
|
||||||
|
):
|
||||||
|
playready_cdm = self.get_cdm(self.service, self.profile, drm="playready")
|
||||||
|
if playready_cdm:
|
||||||
|
self.log.info("Switching to PlayReady CDM for PlayReady content")
|
||||||
|
self.cdm = playready_cdm
|
||||||
|
|
||||||
if isinstance(drm, Widevine):
|
if isinstance(drm, Widevine):
|
||||||
with self.DRM_TABLE_LOCK:
|
with self.DRM_TABLE_LOCK:
|
||||||
cek_tree = Tree(Text.assemble(("Widevine", "cyan"), (f"({drm.pssh.dumps()})", "text"), overflow="fold"))
|
pssh_display = self._truncate_pssh_for_display(drm.pssh.dumps(), "Widevine")
|
||||||
|
cek_tree = Tree(Text.assemble(("Widevine", "cyan"), (f"({pssh_display})", "text"), overflow="fold"))
|
||||||
pre_existing_tree = next(
|
pre_existing_tree = next(
|
||||||
(x for x in table.columns[0].cells if isinstance(x, Tree) and x.label == cek_tree.label), None
|
(x for x in table.columns[0].cells if isinstance(x, Tree) and x.label == cek_tree.label), None
|
||||||
)
|
)
|
||||||
@@ -1300,10 +1333,11 @@ class dl:
|
|||||||
|
|
||||||
elif isinstance(drm, PlayReady):
|
elif isinstance(drm, PlayReady):
|
||||||
with self.DRM_TABLE_LOCK:
|
with self.DRM_TABLE_LOCK:
|
||||||
|
pssh_display = self._truncate_pssh_for_display(drm.pssh_b64 or "", "PlayReady")
|
||||||
cek_tree = Tree(
|
cek_tree = Tree(
|
||||||
Text.assemble(
|
Text.assemble(
|
||||||
("PlayReady", "cyan"),
|
("PlayReady", "cyan"),
|
||||||
(f"({drm.pssh_b64 or ''})", "text"),
|
(f"({pssh_display})", "text"),
|
||||||
overflow="fold",
|
overflow="fold",
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
@@ -1477,26 +1511,17 @@ class dl:
|
|||||||
|
|
||||||
cdm_api = next(iter(x for x in config.remote_cdm if x["name"] == cdm_name), None)
|
cdm_api = next(iter(x for x in config.remote_cdm if x["name"] == cdm_name), None)
|
||||||
if cdm_api:
|
if cdm_api:
|
||||||
is_decrypt_lab = True if cdm_api["type"] == "decrypt_labs" else False
|
is_decrypt_lab = True if cdm_api.get("type") == "decrypt_labs" else False
|
||||||
if is_decrypt_lab:
|
if is_decrypt_lab:
|
||||||
device_type = cdm_api.get("device_type")
|
|
||||||
del cdm_api["name"]
|
del cdm_api["name"]
|
||||||
del cdm_api["type"]
|
del cdm_api["type"]
|
||||||
|
|
||||||
# Use the appropriate DecryptLabs CDM class based on device type
|
# All DecryptLabs CDMs use DecryptLabsRemoteCDM
|
||||||
if device_type == "PLAYREADY" or cdm_api.get("device_name") in ["SL2", "SL3"]:
|
return DecryptLabsRemoteCDM(service_name=service, vaults=self.vaults, **cdm_api)
|
||||||
from unshackle.core.cdm.decrypt_labs_remote_cdm import DecryptLabsRemotePlayReadyCDM
|
|
||||||
|
|
||||||
# Remove unused parameters for PlayReady CDM
|
|
||||||
cdm_params = cdm_api.copy()
|
|
||||||
cdm_params.pop("device_type", None)
|
|
||||||
cdm_params.pop("system_id", None)
|
|
||||||
return DecryptLabsRemotePlayReadyCDM(service_name=service, vaults=self.vaults, **cdm_params)
|
|
||||||
else:
|
|
||||||
return DecryptLabsRemoteCDM(service_name=service, vaults=self.vaults, **cdm_api)
|
|
||||||
else:
|
else:
|
||||||
del cdm_api["name"]
|
del cdm_api["name"]
|
||||||
del cdm_api["type"]
|
if "type" in cdm_api:
|
||||||
|
del cdm_api["type"]
|
||||||
return RemoteCdm(**cdm_api)
|
return RemoteCdm(**cdm_api)
|
||||||
|
|
||||||
prd_path = config.directories.prds / f"{cdm_name}.prd"
|
prd_path = config.directories.prds / f"{cdm_name}.prd"
|
||||||
|
|||||||
@@ -12,84 +12,113 @@ from unshackle.core.vault import Vault
|
|||||||
from unshackle.core.vaults import Vaults
|
from unshackle.core.vaults import Vaults
|
||||||
|
|
||||||
|
|
||||||
|
def _load_vaults(vault_names: list[str]) -> Vaults:
|
||||||
|
"""Load and validate vaults by name."""
|
||||||
|
vaults = Vaults()
|
||||||
|
for vault_name in vault_names:
|
||||||
|
vault_config = next((x for x in config.key_vaults if x["name"] == vault_name), None)
|
||||||
|
if not vault_config:
|
||||||
|
raise click.ClickException(f"Vault ({vault_name}) is not defined in the config.")
|
||||||
|
|
||||||
|
vault_type = vault_config["type"]
|
||||||
|
vault_args = vault_config.copy()
|
||||||
|
del vault_args["type"]
|
||||||
|
|
||||||
|
if not vaults.load(vault_type, **vault_args):
|
||||||
|
raise click.ClickException(f"Failed to load vault ({vault_name}).")
|
||||||
|
|
||||||
|
return vaults
|
||||||
|
|
||||||
|
|
||||||
|
def _process_service_keys(from_vault: Vault, service: str, log: logging.Logger) -> dict[str, str]:
|
||||||
|
"""Get and validate keys from a vault for a specific service."""
|
||||||
|
content_keys = list(from_vault.get_keys(service))
|
||||||
|
|
||||||
|
bad_keys = {kid: key for kid, key in content_keys if not key or key.count("0") == len(key)}
|
||||||
|
for kid, key in bad_keys.items():
|
||||||
|
log.warning(f"Skipping NULL key: {kid}:{key}")
|
||||||
|
|
||||||
|
return {kid: key for kid, key in content_keys if kid not in bad_keys}
|
||||||
|
|
||||||
|
|
||||||
|
def _copy_service_data(to_vault: Vault, from_vault: Vault, service: str, log: logging.Logger) -> int:
|
||||||
|
"""Copy data for a single service between vaults."""
|
||||||
|
content_keys = _process_service_keys(from_vault, service, log)
|
||||||
|
total_count = len(content_keys)
|
||||||
|
|
||||||
|
if total_count == 0:
|
||||||
|
log.info(f"{service}: No keys found in {from_vault}")
|
||||||
|
return 0
|
||||||
|
|
||||||
|
try:
|
||||||
|
added = to_vault.add_keys(service, content_keys)
|
||||||
|
except PermissionError:
|
||||||
|
log.warning(f"{service}: No permission to create table in {to_vault}, skipped")
|
||||||
|
return 0
|
||||||
|
|
||||||
|
existed = total_count - added
|
||||||
|
|
||||||
|
if added > 0 and existed > 0:
|
||||||
|
log.info(f"{service}: {added} added, {existed} skipped ({total_count} total)")
|
||||||
|
elif added > 0:
|
||||||
|
log.info(f"{service}: {added} added ({total_count} total)")
|
||||||
|
else:
|
||||||
|
log.info(f"{service}: {existed} skipped (all existed)")
|
||||||
|
|
||||||
|
return added
|
||||||
|
|
||||||
|
|
||||||
@click.group(short_help="Manage and configure Key Vaults.", context_settings=context_settings)
|
@click.group(short_help="Manage and configure Key Vaults.", context_settings=context_settings)
|
||||||
def kv() -> None:
|
def kv() -> None:
|
||||||
"""Manage and configure Key Vaults."""
|
"""Manage and configure Key Vaults."""
|
||||||
|
|
||||||
|
|
||||||
@kv.command()
|
@kv.command()
|
||||||
@click.argument("to_vault", type=str)
|
@click.argument("to_vault_name", type=str)
|
||||||
@click.argument("from_vaults", nargs=-1, type=click.UNPROCESSED)
|
@click.argument("from_vault_names", nargs=-1, type=click.UNPROCESSED)
|
||||||
@click.option("-s", "--service", type=str, default=None, help="Only copy data to and from a specific service.")
|
@click.option("-s", "--service", type=str, default=None, help="Only copy data to and from a specific service.")
|
||||||
def copy(to_vault: str, from_vaults: list[str], service: Optional[str] = None) -> None:
|
def copy(to_vault_name: str, from_vault_names: list[str], service: Optional[str] = None) -> None:
|
||||||
"""
|
"""
|
||||||
Copy data from multiple Key Vaults into a single Key Vault.
|
Copy data from multiple Key Vaults into a single Key Vault.
|
||||||
Rows with matching KIDs are skipped unless there's no KEY set.
|
Rows with matching KIDs are skipped unless there's no KEY set.
|
||||||
Existing data is not deleted or altered.
|
Existing data is not deleted or altered.
|
||||||
|
|
||||||
The `to_vault` argument is the key vault you wish to copy data to.
|
The `to_vault_name` argument is the key vault you wish to copy data to.
|
||||||
It should be the name of a Key Vault defined in the config.
|
It should be the name of a Key Vault defined in the config.
|
||||||
|
|
||||||
The `from_vaults` argument is the key vault(s) you wish to take
|
The `from_vault_names` argument is the key vault(s) you wish to take
|
||||||
data from. You may supply multiple key vaults.
|
data from. You may supply multiple key vaults.
|
||||||
"""
|
"""
|
||||||
if not from_vaults:
|
if not from_vault_names:
|
||||||
raise click.ClickException("No Vaults were specified to copy data from.")
|
raise click.ClickException("No Vaults were specified to copy data from.")
|
||||||
|
|
||||||
log = logging.getLogger("kv")
|
log = logging.getLogger("kv")
|
||||||
|
|
||||||
vaults = Vaults()
|
all_vault_names = [to_vault_name] + list(from_vault_names)
|
||||||
for vault_name in [to_vault] + list(from_vaults):
|
vaults = _load_vaults(all_vault_names)
|
||||||
vault = next((x for x in config.key_vaults if x["name"] == vault_name), None)
|
|
||||||
if not vault:
|
|
||||||
raise click.ClickException(f"Vault ({vault_name}) is not defined in the config.")
|
|
||||||
vault_type = vault["type"]
|
|
||||||
vault_args = vault.copy()
|
|
||||||
del vault_args["type"]
|
|
||||||
if not vaults.load(vault_type, **vault_args):
|
|
||||||
raise click.ClickException(f"Failed to load vault ({vault_name}).")
|
|
||||||
|
|
||||||
to_vault: Vault = vaults.vaults[0]
|
to_vault = vaults.vaults[0]
|
||||||
from_vaults: list[Vault] = vaults.vaults[1:]
|
from_vaults = vaults.vaults[1:]
|
||||||
|
|
||||||
|
vault_names = ", ".join([v.name for v in from_vaults])
|
||||||
|
log.info(f"Copying data from {vault_names} → {to_vault.name}")
|
||||||
|
|
||||||
log.info(f"Copying data from {', '.join([x.name for x in from_vaults])}, into {to_vault.name}")
|
|
||||||
if service:
|
if service:
|
||||||
service = Services.get_tag(service)
|
service = Services.get_tag(service)
|
||||||
log.info(f"Only copying data for service {service}")
|
log.info(f"Filtering by service: {service}")
|
||||||
|
|
||||||
total_added = 0
|
total_added = 0
|
||||||
for from_vault in from_vaults:
|
for from_vault in from_vaults:
|
||||||
if service:
|
services_to_copy = [service] if service else from_vault.get_services()
|
||||||
services = [service]
|
|
||||||
else:
|
|
||||||
services = from_vault.get_services()
|
|
||||||
|
|
||||||
for service_ in services:
|
|
||||||
log.info(f"Getting data from {from_vault} for {service_}")
|
|
||||||
content_keys = list(from_vault.get_keys(service_)) # important as it's a generator we iterate twice
|
|
||||||
|
|
||||||
bad_keys = {kid: key for kid, key in content_keys if not key or key.count("0") == len(key)}
|
|
||||||
|
|
||||||
for kid, key in bad_keys.items():
|
|
||||||
log.warning(f"Cannot add a NULL Content Key to a Vault, skipping: {kid}:{key}")
|
|
||||||
|
|
||||||
content_keys = {kid: key for kid, key in content_keys if kid not in bad_keys}
|
|
||||||
|
|
||||||
total_count = len(content_keys)
|
|
||||||
log.info(f"Adding {total_count} Content Keys to {to_vault} for {service_}")
|
|
||||||
|
|
||||||
try:
|
|
||||||
added = to_vault.add_keys(service_, content_keys)
|
|
||||||
except PermissionError:
|
|
||||||
log.warning(f" - No permission to create table ({service_}) in {to_vault}, skipping...")
|
|
||||||
continue
|
|
||||||
|
|
||||||
|
for service_tag in services_to_copy:
|
||||||
|
added = _copy_service_data(to_vault, from_vault, service_tag, log)
|
||||||
total_added += added
|
total_added += added
|
||||||
existed = total_count - added
|
|
||||||
|
|
||||||
log.info(f"{to_vault} ({service_}): {added} newly added, {existed} already existed (skipped)")
|
if total_added > 0:
|
||||||
|
log.info(f"Successfully added {total_added} new keys to {to_vault}")
|
||||||
log.info(f"{to_vault}: {total_added} total newly added")
|
else:
|
||||||
|
log.info("Copy completed - no new keys to add")
|
||||||
|
|
||||||
|
|
||||||
@kv.command()
|
@kv.command()
|
||||||
@@ -106,9 +135,9 @@ def sync(ctx: click.Context, vaults: list[str], service: Optional[str] = None) -
|
|||||||
if not len(vaults) > 1:
|
if not len(vaults) > 1:
|
||||||
raise click.ClickException("You must provide more than one Vault to sync.")
|
raise click.ClickException("You must provide more than one Vault to sync.")
|
||||||
|
|
||||||
ctx.invoke(copy, to_vault=vaults[0], from_vaults=vaults[1:], service=service)
|
ctx.invoke(copy, to_vault_name=vaults[0], from_vault_names=vaults[1:], service=service)
|
||||||
for i in range(1, len(vaults)):
|
for i in range(1, len(vaults)):
|
||||||
ctx.invoke(copy, to_vault=vaults[i], from_vaults=[vaults[i - 1]], service=service)
|
ctx.invoke(copy, to_vault_name=vaults[i], from_vault_names=[vaults[i - 1]], service=service)
|
||||||
|
|
||||||
|
|
||||||
@kv.command()
|
@kv.command()
|
||||||
@@ -135,15 +164,7 @@ def add(file: Path, service: str, vaults: list[str]) -> None:
|
|||||||
log = logging.getLogger("kv")
|
log = logging.getLogger("kv")
|
||||||
service = Services.get_tag(service)
|
service = Services.get_tag(service)
|
||||||
|
|
||||||
vaults_ = Vaults()
|
vaults_ = _load_vaults(list(vaults))
|
||||||
for vault_name in vaults:
|
|
||||||
vault = next((x for x in config.key_vaults if x["name"] == vault_name), None)
|
|
||||||
if not vault:
|
|
||||||
raise click.ClickException(f"Vault ({vault_name}) is not defined in the config.")
|
|
||||||
vault_type = vault["type"]
|
|
||||||
vault_args = vault.copy()
|
|
||||||
del vault_args["type"]
|
|
||||||
vaults_.load(vault_type, **vault_args)
|
|
||||||
|
|
||||||
data = file.read_text(encoding="utf8")
|
data = file.read_text(encoding="utf8")
|
||||||
kid_keys: dict[str, str] = {}
|
kid_keys: dict[str, str] = {}
|
||||||
@@ -173,15 +194,7 @@ def prepare(vaults: list[str]) -> None:
|
|||||||
"""Create Service Tables on Vaults if not yet created."""
|
"""Create Service Tables on Vaults if not yet created."""
|
||||||
log = logging.getLogger("kv")
|
log = logging.getLogger("kv")
|
||||||
|
|
||||||
vaults_ = Vaults()
|
vaults_ = _load_vaults(vaults)
|
||||||
for vault_name in vaults:
|
|
||||||
vault = next((x for x in config.key_vaults if x["name"] == vault_name), None)
|
|
||||||
if not vault:
|
|
||||||
raise click.ClickException(f"Vault ({vault_name}) is not defined in the config.")
|
|
||||||
vault_type = vault["type"]
|
|
||||||
vault_args = vault.copy()
|
|
||||||
del vault_args["type"]
|
|
||||||
vaults_.load(vault_type, **vault_args)
|
|
||||||
|
|
||||||
for vault in vaults_:
|
for vault in vaults_:
|
||||||
if hasattr(vault, "has_table") and hasattr(vault, "create_table"):
|
if hasattr(vault, "has_table") and hasattr(vault, "create_table"):
|
||||||
|
|||||||
@@ -1 +1 @@
|
|||||||
__version__ = "1.4.3"
|
__version__ = "1.4.5"
|
||||||
|
|||||||
File diff suppressed because it is too large
Load Diff
@@ -256,44 +256,38 @@ class PlayReady:
|
|||||||
return keys
|
return keys
|
||||||
|
|
||||||
def get_content_keys(self, cdm: PlayReadyCdm, certificate: Callable, licence: Callable) -> None:
|
def get_content_keys(self, cdm: PlayReadyCdm, certificate: Callable, licence: Callable) -> None:
|
||||||
for kid in self.kids:
|
session_id = cdm.open()
|
||||||
if kid in self.content_keys:
|
try:
|
||||||
continue
|
if hasattr(cdm, "set_pssh_b64") and self.pssh_b64:
|
||||||
|
cdm.set_pssh_b64(self.pssh_b64)
|
||||||
|
|
||||||
session_id = cdm.open()
|
if hasattr(cdm, "set_required_kids"):
|
||||||
try:
|
cdm.set_required_kids(self.kids)
|
||||||
if hasattr(cdm, "set_pssh_b64") and self.pssh_b64:
|
|
||||||
cdm.set_pssh_b64(self.pssh_b64)
|
|
||||||
|
|
||||||
challenge = cdm.get_license_challenge(session_id, self.pssh.wrm_headers[0])
|
challenge = cdm.get_license_challenge(session_id, self.pssh.wrm_headers[0])
|
||||||
|
|
||||||
|
if challenge:
|
||||||
try:
|
try:
|
||||||
license_res = licence(challenge=challenge)
|
license_res = licence(challenge=challenge)
|
||||||
except Exception:
|
if isinstance(license_res, bytes):
|
||||||
if hasattr(cdm, "use_cached_keys_as_fallback"):
|
license_str = license_res.decode(errors="ignore")
|
||||||
if cdm.use_cached_keys_as_fallback(session_id):
|
else:
|
||||||
keys = self._extract_keys_from_cdm(cdm, session_id)
|
license_str = str(license_res)
|
||||||
self.content_keys.update(keys)
|
|
||||||
continue
|
|
||||||
|
|
||||||
|
if "<License>" not in license_str:
|
||||||
|
try:
|
||||||
|
license_str = base64.b64decode(license_str + "===").decode()
|
||||||
|
except Exception:
|
||||||
|
pass
|
||||||
|
|
||||||
|
cdm.parse_license(session_id, license_str)
|
||||||
|
except Exception:
|
||||||
raise
|
raise
|
||||||
|
|
||||||
if isinstance(license_res, bytes):
|
keys = self._extract_keys_from_cdm(cdm, session_id)
|
||||||
license_str = license_res.decode(errors="ignore")
|
self.content_keys.update(keys)
|
||||||
else:
|
finally:
|
||||||
license_str = str(license_res)
|
cdm.close(session_id)
|
||||||
|
|
||||||
if "<License>" not in license_str:
|
|
||||||
try:
|
|
||||||
license_str = base64.b64decode(license_str + "===").decode()
|
|
||||||
except Exception:
|
|
||||||
pass
|
|
||||||
|
|
||||||
cdm.parse_license(session_id, license_str)
|
|
||||||
keys = self._extract_keys_from_cdm(cdm, session_id)
|
|
||||||
self.content_keys.update(keys)
|
|
||||||
finally:
|
|
||||||
cdm.close(session_id)
|
|
||||||
|
|
||||||
if not self.content_keys:
|
if not self.content_keys:
|
||||||
raise PlayReady.Exceptions.EmptyLicense("No Content Keys were within the License")
|
raise PlayReady.Exceptions.EmptyLicense("No Content Keys were within the License")
|
||||||
|
|||||||
@@ -185,6 +185,9 @@ class Widevine:
|
|||||||
if cert and hasattr(cdm, "set_service_certificate"):
|
if cert and hasattr(cdm, "set_service_certificate"):
|
||||||
cdm.set_service_certificate(session_id, cert)
|
cdm.set_service_certificate(session_id, cert)
|
||||||
|
|
||||||
|
if hasattr(cdm, "set_required_kids"):
|
||||||
|
cdm.set_required_kids(self.kids)
|
||||||
|
|
||||||
challenge = cdm.get_license_challenge(session_id, self.pssh)
|
challenge = cdm.get_license_challenge(session_id, self.pssh)
|
||||||
|
|
||||||
if hasattr(cdm, "has_cached_keys") and cdm.has_cached_keys(session_id):
|
if hasattr(cdm, "has_cached_keys") and cdm.has_cached_keys(session_id):
|
||||||
@@ -218,6 +221,9 @@ class Widevine:
|
|||||||
if cert and hasattr(cdm, "set_service_certificate"):
|
if cert and hasattr(cdm, "set_service_certificate"):
|
||||||
cdm.set_service_certificate(session_id, cert)
|
cdm.set_service_certificate(session_id, cert)
|
||||||
|
|
||||||
|
if hasattr(cdm, "set_required_kids"):
|
||||||
|
cdm.set_required_kids(self.kids)
|
||||||
|
|
||||||
challenge = cdm.get_license_challenge(session_id, self.pssh)
|
challenge = cdm.get_license_challenge(session_id, self.pssh)
|
||||||
|
|
||||||
if hasattr(cdm, "has_cached_keys") and cdm.has_cached_keys(session_id):
|
if hasattr(cdm, "has_cached_keys") and cdm.has_cached_keys(session_id):
|
||||||
|
|||||||
@@ -420,6 +420,15 @@ class Track:
|
|||||||
for drm in self.drm:
|
for drm in self.drm:
|
||||||
if isinstance(drm, PlayReady):
|
if isinstance(drm, PlayReady):
|
||||||
return drm
|
return drm
|
||||||
|
elif hasattr(cdm, 'is_playready'):
|
||||||
|
if cdm.is_playready:
|
||||||
|
for drm in self.drm:
|
||||||
|
if isinstance(drm, PlayReady):
|
||||||
|
return drm
|
||||||
|
else:
|
||||||
|
for drm in self.drm:
|
||||||
|
if isinstance(drm, Widevine):
|
||||||
|
return drm
|
||||||
|
|
||||||
return self.drm[0]
|
return self.drm[0]
|
||||||
|
|
||||||
|
|||||||
@@ -8,6 +8,7 @@ import tempfile
|
|||||||
from difflib import SequenceMatcher
|
from difflib import SequenceMatcher
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
from typing import Optional, Tuple
|
from typing import Optional, Tuple
|
||||||
|
from xml.sax.saxutils import escape
|
||||||
|
|
||||||
import requests
|
import requests
|
||||||
from requests.adapters import HTTPAdapter, Retry
|
from requests.adapters import HTTPAdapter, Retry
|
||||||
@@ -289,9 +290,9 @@ def _apply_tags(path: Path, tags: dict[str, str]) -> None:
|
|||||||
log.debug("mkvpropedit not found on PATH; skipping tags")
|
log.debug("mkvpropedit not found on PATH; skipping tags")
|
||||||
return
|
return
|
||||||
log.debug("Applying tags to %s: %s", path, tags)
|
log.debug("Applying tags to %s: %s", path, tags)
|
||||||
xml_lines = ["<?xml version='1.0' encoding='UTF-8'?>", "<Tags>", " <Tag>", " <Targets/>"]
|
xml_lines = ['<?xml version="1.0" encoding="UTF-8"?>', "<Tags>", " <Tag>", " <Targets/>"]
|
||||||
for name, value in tags.items():
|
for name, value in tags.items():
|
||||||
xml_lines.append(f" <Simple><Name>{name}</Name><String>{value}</String></Simple>")
|
xml_lines.append(f" <Simple><Name>{escape(name)}</Name><String>{escape(value)}</String></Simple>")
|
||||||
xml_lines.extend([" </Tag>", "</Tags>"])
|
xml_lines.extend([" </Tag>", "</Tags>"])
|
||||||
with tempfile.NamedTemporaryFile("w", suffix=".xml", delete=False) as f:
|
with tempfile.NamedTemporaryFile("w", suffix=".xml", delete=False) as f:
|
||||||
f.write("\n".join(xml_lines))
|
f.write("\n".join(xml_lines))
|
||||||
@@ -349,13 +350,25 @@ def tag_file(path: Path, title: Title, tmdb_id: Optional[int] | None = None) ->
|
|||||||
if simkl_tmdb_id:
|
if simkl_tmdb_id:
|
||||||
tmdb_id = simkl_tmdb_id
|
tmdb_id = simkl_tmdb_id
|
||||||
|
|
||||||
show_ids = simkl_data.get("show", {}).get("ids", {})
|
# Handle TV show data from Simkl
|
||||||
if show_ids.get("imdb"):
|
if simkl_data.get("type") == "episode" and "show" in simkl_data:
|
||||||
standard_tags["IMDB"] = f"https://www.imdb.com/title/{show_ids['imdb']}"
|
show_ids = simkl_data.get("show", {}).get("ids", {})
|
||||||
if show_ids.get("tvdb"):
|
if show_ids.get("imdb"):
|
||||||
standard_tags["TVDB"] = f"https://thetvdb.com/dereferrer/series/{show_ids['tvdb']}"
|
standard_tags["IMDB"] = show_ids["imdb"]
|
||||||
if show_ids.get("tmdbtv"):
|
if show_ids.get("tvdb"):
|
||||||
standard_tags["TMDB"] = f"https://www.themoviedb.org/tv/{show_ids['tmdbtv']}"
|
standard_tags["TVDB2"] = f"series/{show_ids['tvdb']}"
|
||||||
|
if show_ids.get("tmdbtv"):
|
||||||
|
standard_tags["TMDB"] = f"tv/{show_ids['tmdbtv']}"
|
||||||
|
|
||||||
|
# Handle movie data from Simkl
|
||||||
|
elif simkl_data.get("type") == "movie" and "movie" in simkl_data:
|
||||||
|
movie_ids = simkl_data.get("movie", {}).get("ids", {})
|
||||||
|
if movie_ids.get("imdb"):
|
||||||
|
standard_tags["IMDB"] = movie_ids["imdb"]
|
||||||
|
if movie_ids.get("tvdb"):
|
||||||
|
standard_tags["TVDB2"] = f"movies/{movie_ids['tvdb']}"
|
||||||
|
if movie_ids.get("tmdb"):
|
||||||
|
standard_tags["TMDB"] = f"movie/{movie_ids['tmdb']}"
|
||||||
|
|
||||||
# Use TMDB API for additional metadata (either from provided ID or Simkl lookup)
|
# Use TMDB API for additional metadata (either from provided ID or Simkl lookup)
|
||||||
api_key = _api_key()
|
api_key = _api_key()
|
||||||
@@ -373,8 +386,8 @@ def tag_file(path: Path, title: Title, tmdb_id: Optional[int] | None = None) ->
|
|||||||
_apply_tags(path, custom_tags)
|
_apply_tags(path, custom_tags)
|
||||||
return
|
return
|
||||||
|
|
||||||
tmdb_url = f"https://www.themoviedb.org/{'movie' if kind == 'movie' else 'tv'}/{tmdb_id}"
|
prefix = "movie" if kind == "movie" else "tv"
|
||||||
standard_tags["TMDB"] = tmdb_url
|
standard_tags["TMDB"] = f"{prefix}/{tmdb_id}"
|
||||||
try:
|
try:
|
||||||
ids = external_ids(tmdb_id, kind)
|
ids = external_ids(tmdb_id, kind)
|
||||||
except requests.RequestException as exc:
|
except requests.RequestException as exc:
|
||||||
@@ -385,11 +398,13 @@ def tag_file(path: Path, title: Title, tmdb_id: Optional[int] | None = None) ->
|
|||||||
|
|
||||||
imdb_id = ids.get("imdb_id")
|
imdb_id = ids.get("imdb_id")
|
||||||
if imdb_id:
|
if imdb_id:
|
||||||
standard_tags["IMDB"] = f"https://www.imdb.com/title/{imdb_id}"
|
standard_tags["IMDB"] = imdb_id
|
||||||
tvdb_id = ids.get("tvdb_id")
|
tvdb_id = ids.get("tvdb_id")
|
||||||
if tvdb_id:
|
if tvdb_id:
|
||||||
tvdb_prefix = "movies" if kind == "movie" else "series"
|
if kind == "movie":
|
||||||
standard_tags["TVDB"] = f"https://thetvdb.com/dereferrer/{tvdb_prefix}/{tvdb_id}"
|
standard_tags["TVDB2"] = f"movies/{tvdb_id}"
|
||||||
|
else:
|
||||||
|
standard_tags["TVDB2"] = f"series/{tvdb_id}"
|
||||||
|
|
||||||
merged_tags = {
|
merged_tags = {
|
||||||
**custom_tags,
|
**custom_tags,
|
||||||
|
|||||||
@@ -282,6 +282,10 @@ class EXAMPLE(Service):
|
|||||||
|
|
||||||
return chapters
|
return chapters
|
||||||
|
|
||||||
|
def get_widevine_service_certificate(self, **_: any) -> str:
|
||||||
|
"""Return the Widevine service certificate from config, if available."""
|
||||||
|
return self.config.get("certificate")
|
||||||
|
|
||||||
def get_playready_license(self, *, challenge: bytes, title: Title_T, track: AnyTrack) -> Optional[bytes]:
|
def get_playready_license(self, *, challenge: bytes, title: Title_T, track: AnyTrack) -> Optional[bytes]:
|
||||||
"""Retrieve a PlayReady license for a given track."""
|
"""Retrieve a PlayReady license for a given track."""
|
||||||
|
|
||||||
|
|||||||
@@ -117,7 +117,7 @@ remote_cdm:
|
|||||||
type: "decrypt_labs"
|
type: "decrypt_labs"
|
||||||
device_name: "L1" # Scheme identifier - must match exactly
|
device_name: "L1" # Scheme identifier - must match exactly
|
||||||
device_type: ANDROID
|
device_type: ANDROID
|
||||||
system_id: 4464
|
system_id: 4464
|
||||||
security_level: 1
|
security_level: 1
|
||||||
host: "https://keyxtractor.decryptlabs.com"
|
host: "https://keyxtractor.decryptlabs.com"
|
||||||
secret: "your_decrypt_labs_api_key_here"
|
secret: "your_decrypt_labs_api_key_here"
|
||||||
|
|||||||
@@ -131,16 +131,27 @@ class MySQL(Vault):
|
|||||||
if any(isinstance(kid, UUID) for kid, key_ in kid_keys.items()):
|
if any(isinstance(kid, UUID) for kid, key_ in kid_keys.items()):
|
||||||
kid_keys = {kid.hex if isinstance(kid, UUID) else kid: key_ for kid, key_ in kid_keys.items()}
|
kid_keys = {kid.hex if isinstance(kid, UUID) else kid: key_ for kid, key_ in kid_keys.items()}
|
||||||
|
|
||||||
|
if not kid_keys:
|
||||||
|
return 0
|
||||||
|
|
||||||
conn = self.conn_factory.get()
|
conn = self.conn_factory.get()
|
||||||
cursor = conn.cursor()
|
cursor = conn.cursor()
|
||||||
|
|
||||||
try:
|
try:
|
||||||
|
placeholders = ",".join(["%s"] * len(kid_keys))
|
||||||
|
cursor.execute(f"SELECT kid FROM `{service}` WHERE kid IN ({placeholders})", list(kid_keys.keys()))
|
||||||
|
existing_kids = {row["kid"] for row in cursor.fetchall()}
|
||||||
|
|
||||||
|
new_keys = {kid: key for kid, key in kid_keys.items() if kid not in existing_kids}
|
||||||
|
|
||||||
|
if not new_keys:
|
||||||
|
return 0
|
||||||
|
|
||||||
cursor.executemany(
|
cursor.executemany(
|
||||||
# TODO: SQL injection risk
|
f"INSERT INTO `{service}` (kid, key_) VALUES (%s, %s)",
|
||||||
f"INSERT IGNORE INTO `{service}` (kid, key_) VALUES (%s, %s)",
|
new_keys.items(),
|
||||||
kid_keys.items(),
|
|
||||||
)
|
)
|
||||||
return cursor.rowcount
|
return len(new_keys)
|
||||||
finally:
|
finally:
|
||||||
conn.commit()
|
conn.commit()
|
||||||
cursor.close()
|
cursor.close()
|
||||||
|
|||||||
@@ -102,16 +102,27 @@ class SQLite(Vault):
|
|||||||
if any(isinstance(kid, UUID) for kid, key_ in kid_keys.items()):
|
if any(isinstance(kid, UUID) for kid, key_ in kid_keys.items()):
|
||||||
kid_keys = {kid.hex if isinstance(kid, UUID) else kid: key_ for kid, key_ in kid_keys.items()}
|
kid_keys = {kid.hex if isinstance(kid, UUID) else kid: key_ for kid, key_ in kid_keys.items()}
|
||||||
|
|
||||||
|
if not kid_keys:
|
||||||
|
return 0
|
||||||
|
|
||||||
conn = self.conn_factory.get()
|
conn = self.conn_factory.get()
|
||||||
cursor = conn.cursor()
|
cursor = conn.cursor()
|
||||||
|
|
||||||
try:
|
try:
|
||||||
|
placeholders = ",".join(["?"] * len(kid_keys))
|
||||||
|
cursor.execute(f"SELECT kid FROM `{service}` WHERE kid IN ({placeholders})", list(kid_keys.keys()))
|
||||||
|
existing_kids = {row[0] for row in cursor.fetchall()}
|
||||||
|
|
||||||
|
new_keys = {kid: key for kid, key in kid_keys.items() if kid not in existing_kids}
|
||||||
|
|
||||||
|
if not new_keys:
|
||||||
|
return 0
|
||||||
|
|
||||||
cursor.executemany(
|
cursor.executemany(
|
||||||
# TODO: SQL injection risk
|
f"INSERT INTO `{service}` (kid, key_) VALUES (?, ?)",
|
||||||
f"INSERT OR IGNORE INTO `{service}` (kid, key_) VALUES (?, ?)",
|
new_keys.items(),
|
||||||
kid_keys.items(),
|
|
||||||
)
|
)
|
||||||
return cursor.rowcount
|
return len(new_keys)
|
||||||
finally:
|
finally:
|
||||||
conn.commit()
|
conn.commit()
|
||||||
cursor.close()
|
cursor.close()
|
||||||
|
|||||||
Reference in New Issue
Block a user