mirror of
https://github.com/unshackle-dl/unshackle.git
synced 2025-10-23 15:11:08 +00:00
Compare commits
36 Commits
b4efdf3f2c
...
1.4.7
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
bc26bf3046 | ||
|
|
35efdbff6d | ||
|
|
63b7a49c1a | ||
|
|
98ecf6f876 | ||
|
|
5df6914536 | ||
|
|
c1df074965 | ||
|
|
da60a396dd | ||
|
|
a99a391395 | ||
|
|
ed32939d83 | ||
|
|
4006593a8a | ||
|
|
307be4549b | ||
|
|
a82828768d | ||
|
|
d18a5de0d0 | ||
|
|
04b540b363 | ||
|
|
6137146705 | ||
|
|
859d09693c | ||
|
|
5f022635cb | ||
|
|
ad66502c0c | ||
|
|
e462f07b7a | ||
|
|
83b600e999 | ||
|
|
ea8a7b00c9 | ||
|
|
16ee4175a4 | ||
|
|
f722ec69b6 | ||
|
|
2330297ea4 | ||
|
|
86bb162868 | ||
|
|
501cfd68e8 | ||
|
|
76fb2eea95 | ||
|
|
ea5ec40bcd | ||
|
|
329850b043 | ||
|
|
73595f3b50 | ||
|
|
1e82283133 | ||
|
|
ab13dde9d2 | ||
|
|
9fd0895128 | ||
|
|
ed744205ad | ||
|
|
3ef43afeed | ||
|
|
26851cbe7c |
151
CHANGELOG.md
151
CHANGELOG.md
@@ -5,6 +5,157 @@ All notable changes to this project will be documented in this file.
|
|||||||
The format is based on [Keep a Changelog](https://keepachangelog.com/en/1.1.0/),
|
The format is based on [Keep a Changelog](https://keepachangelog.com/en/1.1.0/),
|
||||||
and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0.html).
|
and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0.html).
|
||||||
|
|
||||||
|
## [1.4.7] - 2025-09-25
|
||||||
|
|
||||||
|
### Added
|
||||||
|
|
||||||
|
- **curl_cffi Session Support**: Enhanced anti-bot protection with browser impersonation
|
||||||
|
- Added new session utility with curl_cffi support for bypassing anti-bot measures
|
||||||
|
- Browser impersonation support for Chrome, Firefox, and Safari user agents
|
||||||
|
- Full backward compatibility with requests.Session maintained
|
||||||
|
- Suppressed HTTPS proxy warnings for improved user experience
|
||||||
|
- **Download Retry Functionality**: Configurable retry mechanism for failed downloads
|
||||||
|
- Added retry count option to download function for improved reliability
|
||||||
|
- **Subtitle Requirements Options**: Enhanced subtitle download control
|
||||||
|
- Added options for required subtitles in download command
|
||||||
|
- Better control over subtitle track selection and requirements
|
||||||
|
- **Quality Selection Enhancement**: Improved quality selection options
|
||||||
|
- Added best available quality option in download command for optimal track selection
|
||||||
|
- **DecryptLabs API Integration**: Enhanced remote CDM configuration
|
||||||
|
- Added decrypt_labs_api_key to Config initialization for better API integration
|
||||||
|
|
||||||
|
### Changed
|
||||||
|
|
||||||
|
- **Manifest Parser Updates**: Enhanced compatibility across all parsers
|
||||||
|
- Updated DASH, HLS, ISM, and M3U8 parsers to accept curl_cffi sessions
|
||||||
|
- Improved cookie handling compatibility between requests and curl_cffi
|
||||||
|
- **Logging Improvements**: Reduced log verbosity for better user experience
|
||||||
|
- Changed duplicate track log level to debug to reduce console noise
|
||||||
|
- Dynamic CDM selection messages moved to debug-only output
|
||||||
|
|
||||||
|
### Fixed
|
||||||
|
|
||||||
|
- **Remote CDM Reuse**: Fixed KeyError in dynamic CDM selection
|
||||||
|
- Prevents KeyError when reusing remote CDMs in dynamic selection process
|
||||||
|
- Creates copy of CDM dictionary before modification to prevent configuration mutation
|
||||||
|
- Allows same CDM to be selected multiple times within session without errors
|
||||||
|
|
||||||
|
## [1.4.6] - 2025-09-13
|
||||||
|
|
||||||
|
### Added
|
||||||
|
|
||||||
|
- **Quality-Based CDM Selection**: Dynamic CDM selection based on video resolution
|
||||||
|
- Automatically selects appropriate CDM (L3/L1) based on video track quality
|
||||||
|
- Supports quality thresholds in configuration (>=, >, <=, <, exact match)
|
||||||
|
- Pre-selects optimal CDM based on highest quality across all video tracks
|
||||||
|
- Maintains backward compatibility with existing CDM configurations
|
||||||
|
- **Automatic Audio Language Metadata**: Intelligent embedded audio language detection
|
||||||
|
- Automatically sets audio language metadata when no separate audio tracks exist
|
||||||
|
- Smart video track selection based on title language with fallbacks
|
||||||
|
- Enhanced FFmpeg repackaging with audio stream metadata injection
|
||||||
|
- **Lazy DRM Loading**: Deferred DRM loading for multi-track key retrieval optimization
|
||||||
|
- Add deferred DRM loading to M3U8 parser to mark tracks for later processing
|
||||||
|
- Just-in-time DRM loading during download process for better performance
|
||||||
|
|
||||||
|
### Changed
|
||||||
|
|
||||||
|
- **Enhanced CDM Management**: Improved CDM switching logic for multi-quality downloads
|
||||||
|
- CDM selection now based on highest quality track to avoid inefficient switching
|
||||||
|
- Quality-based selection only within same DRM type (Widevine-to-Widevine, PlayReady-to-PlayReady)
|
||||||
|
- Single CDM used per session for better performance and reliability
|
||||||
|
|
||||||
|
### Fixed
|
||||||
|
|
||||||
|
- **Vault Caching Issues**: Fixed vault count display and NoneType iteration errors
|
||||||
|
- Fix 'NoneType' object is not iterable error in DecryptLabsRemoteCDM
|
||||||
|
- Fix vault count display showing 0/3 instead of actual successful vault count
|
||||||
|
- **Service Name Transmission**: Resolved DecryptLabsRemoteCDM service name issues
|
||||||
|
- Fixed DecryptLabsRemoteCDM sending 'generic' instead of proper service names
|
||||||
|
- Added case-insensitive vault lookups for SQLite/MySQL vaults
|
||||||
|
- Added local vault integration to DecryptLabsRemoteCDM
|
||||||
|
- **Import Organization**: Improved import ordering and code formatting
|
||||||
|
- Reorder imports in decrypt_labs_remote_cdm.py for better organization
|
||||||
|
- Clean up trailing whitespace in vault files
|
||||||
|
|
||||||
|
### Configuration
|
||||||
|
|
||||||
|
- **New CDM Configuration Format**: Extended `cdm:` section supports quality-based selection
|
||||||
|
```yaml
|
||||||
|
cdm:
|
||||||
|
SERVICE_NAME:
|
||||||
|
"<=1080": l3_cdm_name
|
||||||
|
">1080": l1_cdm_name
|
||||||
|
default: l3_cdm_name
|
||||||
|
```
|
||||||
|
|
||||||
|
## [1.4.5] - 2025-09-09
|
||||||
|
|
||||||
|
### Added
|
||||||
|
|
||||||
|
- **Enhanced CDM Key Caching**: Improved key caching and session management for L1/L2 devices
|
||||||
|
- Optimized `get_cached_keys_if_exists` functionality for better performance with L1/L2 devices
|
||||||
|
- Enhanced cached key retrieval logic with improved session handling
|
||||||
|
- **Widevine Common Certificate Fallback**: Added fallback to Widevine common certificate for L1 devices
|
||||||
|
- Improved compatibility for L1 devices when service certificates are unavailable
|
||||||
|
- **Enhanced Vault Loading**: Improved vault loading and key copying logic
|
||||||
|
- Better error handling and key management in vault operations
|
||||||
|
- **PSSH Display Optimization**: Truncated PSSH string display in non-debug mode for cleaner output
|
||||||
|
- **CDM Error Messaging**: Added error messages for missing service certificates in CDM sessions
|
||||||
|
|
||||||
|
### Changed
|
||||||
|
|
||||||
|
- **Dynamic Version Headers**: Updated User-Agent headers to use dynamic version strings
|
||||||
|
- DecryptLabsRemoteCDM now uses dynamic version import instead of hardcoded version
|
||||||
|
- **Intelligent CDM Caching**: Implemented intelligent caching system for CDM license requests
|
||||||
|
- Enhanced caching logic reduces redundant license requests and improves performance
|
||||||
|
- **Enhanced Tag Handling**: Improved tag handling for TV shows and movies from Simkl data
|
||||||
|
- Better metadata processing and formatting for improved media tagging
|
||||||
|
|
||||||
|
### Fixed
|
||||||
|
|
||||||
|
- **CDM Session Management**: Clean up session data when retrieving cached keys
|
||||||
|
- Remove decrypt_labs_session_id and challenge from session when cached keys exist but there are missing kids
|
||||||
|
- Ensures clean state for subsequent requests and prevents session conflicts
|
||||||
|
- **Tag Formatting**: Fixed formatting issues in tag processing
|
||||||
|
- **Import Order**: Fixed import order issues in tags module
|
||||||
|
|
||||||
|
## [1.4.4] - 2025-09-02
|
||||||
|
|
||||||
|
### Added
|
||||||
|
|
||||||
|
- **Enhanced DecryptLabs CDM Support**: Comprehensive remote CDM functionality
|
||||||
|
- Full support for Widevine, PlayReady, and ChromeCDM through DecryptLabsRemoteCDM
|
||||||
|
- Enhanced session management and caching support for remote WV/PR operations
|
||||||
|
- Support for cached keys and improved license handling
|
||||||
|
- New CDM configurations for Chrome and PlayReady devices with updated User-Agent and service certificate
|
||||||
|
- **Advanced Configuration Options**: New device and language preferences
|
||||||
|
- Added configuration options for device certificate status list
|
||||||
|
- Enhanced language preference settings
|
||||||
|
|
||||||
|
### Changed
|
||||||
|
|
||||||
|
- **DRM Decryption Enhancements**: Streamlined decryption process
|
||||||
|
- Simplified decrypt method by removing unused parameter and streamlined logic
|
||||||
|
- Improved DecryptLabs CDM configurations with better device support
|
||||||
|
|
||||||
|
### Fixed
|
||||||
|
|
||||||
|
- **Matroska Tag Compliance**: Enhanced media container compatibility
|
||||||
|
- Fixed Matroska tag compliance with official specification
|
||||||
|
- **Application Branding**: Cleaned up version display
|
||||||
|
- Removed old devine version reference from banner to avoid developer confusion
|
||||||
|
- Updated branding while maintaining original GNU license compliance
|
||||||
|
- **IP Information Handling**: Improved geolocation services
|
||||||
|
- Enhanced get_ip_info functionality with better failover handling
|
||||||
|
- Added support for 429 error handling and multiple API provider fallback
|
||||||
|
- Implemented cached IP info retrieval with fallback tester to avoid rate limiting
|
||||||
|
- **Dependencies**: Streamlined package requirements
|
||||||
|
- Removed unnecessary data extra requirement from langcodes
|
||||||
|
|
||||||
|
### Removed
|
||||||
|
|
||||||
|
- Deprecated version references in application banner for clarity
|
||||||
|
|
||||||
## [1.4.3] - 2025-08-20
|
## [1.4.3] - 2025-08-20
|
||||||
|
|
||||||
### Added
|
### Added
|
||||||
|
|||||||
@@ -4,7 +4,7 @@ build-backend = "hatchling.build"
|
|||||||
|
|
||||||
[project]
|
[project]
|
||||||
name = "unshackle"
|
name = "unshackle"
|
||||||
version = "1.4.3"
|
version = "1.4.6"
|
||||||
description = "Modular Movie, TV, and Music Archival Software."
|
description = "Modular Movie, TV, and Music Archival Software."
|
||||||
authors = [{ name = "unshackle team" }]
|
authors = [{ name = "unshackle team" }]
|
||||||
requires-python = ">=3.10,<3.13"
|
requires-python = ">=3.10,<3.13"
|
||||||
|
|||||||
@@ -66,6 +66,18 @@ from unshackle.core.vaults import Vaults
|
|||||||
|
|
||||||
|
|
||||||
class dl:
|
class dl:
|
||||||
|
@staticmethod
|
||||||
|
def _truncate_pssh_for_display(pssh_string: str, drm_type: str) -> str:
|
||||||
|
"""Truncate PSSH string for display when not in debug mode."""
|
||||||
|
if logging.root.level == logging.DEBUG or not pssh_string:
|
||||||
|
return pssh_string
|
||||||
|
|
||||||
|
max_width = console.width - len(drm_type) - 12
|
||||||
|
if len(pssh_string) <= max_width:
|
||||||
|
return pssh_string
|
||||||
|
|
||||||
|
return pssh_string[: max_width - 3] + "..."
|
||||||
|
|
||||||
@click.command(
|
@click.command(
|
||||||
short_help="Download, Decrypt, and Mux tracks for titles from a Service.",
|
short_help="Download, Decrypt, and Mux tracks for titles from a Service.",
|
||||||
cls=Services,
|
cls=Services,
|
||||||
@@ -161,6 +173,12 @@ class dl:
|
|||||||
help="Language wanted for Audio, overrides -l/--lang for audio tracks.",
|
help="Language wanted for Audio, overrides -l/--lang for audio tracks.",
|
||||||
)
|
)
|
||||||
@click.option("-sl", "--s-lang", type=LANGUAGE_RANGE, default=["all"], help="Language wanted for Subtitles.")
|
@click.option("-sl", "--s-lang", type=LANGUAGE_RANGE, default=["all"], help="Language wanted for Subtitles.")
|
||||||
|
@click.option(
|
||||||
|
"--require-subs",
|
||||||
|
type=LANGUAGE_RANGE,
|
||||||
|
default=[],
|
||||||
|
help="Required subtitle languages. Downloads all subtitles only if these languages exist. Cannot be used with --s-lang.",
|
||||||
|
)
|
||||||
@click.option("-fs", "--forced-subs", is_flag=True, default=False, help="Include forced subtitle tracks.")
|
@click.option("-fs", "--forced-subs", is_flag=True, default=False, help="Include forced subtitle tracks.")
|
||||||
@click.option(
|
@click.option(
|
||||||
"--proxy",
|
"--proxy",
|
||||||
@@ -251,6 +269,13 @@ class dl:
|
|||||||
@click.option(
|
@click.option(
|
||||||
"--reset-cache", "reset_cache", is_flag=True, default=False, help="Clear title cache before fetching."
|
"--reset-cache", "reset_cache", is_flag=True, default=False, help="Clear title cache before fetching."
|
||||||
)
|
)
|
||||||
|
@click.option(
|
||||||
|
"--best-available",
|
||||||
|
"best_available",
|
||||||
|
is_flag=True,
|
||||||
|
default=False,
|
||||||
|
help="Continue with best available quality if requested resolutions are not available.",
|
||||||
|
)
|
||||||
@click.pass_context
|
@click.pass_context
|
||||||
def cli(ctx: click.Context, **kwargs: Any) -> dl:
|
def cli(ctx: click.Context, **kwargs: Any) -> dl:
|
||||||
return dl(ctx, **kwargs)
|
return dl(ctx, **kwargs)
|
||||||
@@ -299,21 +324,6 @@ class dl:
|
|||||||
if getattr(config, "decryption_map", None):
|
if getattr(config, "decryption_map", None):
|
||||||
config.decryption = config.decryption_map.get(self.service, config.decryption)
|
config.decryption = config.decryption_map.get(self.service, config.decryption)
|
||||||
|
|
||||||
with console.status("Loading DRM CDM...", spinner="dots"):
|
|
||||||
try:
|
|
||||||
self.cdm = self.get_cdm(self.service, self.profile)
|
|
||||||
except ValueError as e:
|
|
||||||
self.log.error(f"Failed to load CDM, {e}")
|
|
||||||
sys.exit(1)
|
|
||||||
|
|
||||||
if self.cdm:
|
|
||||||
if hasattr(self.cdm, "device_type") and self.cdm.device_type.name in ["ANDROID", "CHROME"]:
|
|
||||||
self.log.info(f"Loaded Widevine CDM: {self.cdm.system_id} (L{self.cdm.security_level})")
|
|
||||||
else:
|
|
||||||
self.log.info(
|
|
||||||
f"Loaded PlayReady CDM: {self.cdm.certificate_chain.get_name()} (L{self.cdm.security_level})"
|
|
||||||
)
|
|
||||||
|
|
||||||
with console.status("Loading Key Vaults...", spinner="dots"):
|
with console.status("Loading Key Vaults...", spinner="dots"):
|
||||||
self.vaults = Vaults(self.service)
|
self.vaults = Vaults(self.service)
|
||||||
total_vaults = len(config.key_vaults)
|
total_vaults = len(config.key_vaults)
|
||||||
@@ -325,6 +335,16 @@ class dl:
|
|||||||
vault_copy = vault.copy()
|
vault_copy = vault.copy()
|
||||||
del vault_copy["type"]
|
del vault_copy["type"]
|
||||||
|
|
||||||
|
if vault_type.lower() == "api" and "decrypt_labs" in vault_name.lower():
|
||||||
|
if "token" not in vault_copy or not vault_copy["token"]:
|
||||||
|
if config.decrypt_labs_api_key:
|
||||||
|
vault_copy["token"] = config.decrypt_labs_api_key
|
||||||
|
else:
|
||||||
|
self.log.warning(
|
||||||
|
f"No token provided for DecryptLabs vault '{vault_name}' and no global "
|
||||||
|
"decrypt_labs_api_key configured"
|
||||||
|
)
|
||||||
|
|
||||||
if vault_type.lower() == "sqlite":
|
if vault_type.lower() == "sqlite":
|
||||||
try:
|
try:
|
||||||
self.vaults.load_critical(vault_type, **vault_copy)
|
self.vaults.load_critical(vault_type, **vault_copy)
|
||||||
@@ -352,6 +372,24 @@ class dl:
|
|||||||
else:
|
else:
|
||||||
self.log.debug("No vaults are currently active")
|
self.log.debug("No vaults are currently active")
|
||||||
|
|
||||||
|
with console.status("Loading DRM CDM...", spinner="dots"):
|
||||||
|
try:
|
||||||
|
self.cdm = self.get_cdm(self.service, self.profile)
|
||||||
|
except ValueError as e:
|
||||||
|
self.log.error(f"Failed to load CDM, {e}")
|
||||||
|
sys.exit(1)
|
||||||
|
|
||||||
|
if self.cdm:
|
||||||
|
if isinstance(self.cdm, DecryptLabsRemoteCDM):
|
||||||
|
drm_type = "PlayReady" if self.cdm.is_playready else "Widevine"
|
||||||
|
self.log.info(f"Loaded {drm_type} Remote CDM: DecryptLabs (L{self.cdm.security_level})")
|
||||||
|
elif hasattr(self.cdm, "device_type") and self.cdm.device_type.name in ["ANDROID", "CHROME"]:
|
||||||
|
self.log.info(f"Loaded Widevine CDM: {self.cdm.system_id} (L{self.cdm.security_level})")
|
||||||
|
else:
|
||||||
|
self.log.info(
|
||||||
|
f"Loaded PlayReady CDM: {self.cdm.certificate_chain.get_name()} (L{self.cdm.security_level})"
|
||||||
|
)
|
||||||
|
|
||||||
self.proxy_providers = []
|
self.proxy_providers = []
|
||||||
if no_proxy:
|
if no_proxy:
|
||||||
ctx.params["proxy"] = None
|
ctx.params["proxy"] = None
|
||||||
@@ -427,6 +465,7 @@ class dl:
|
|||||||
v_lang: list[str],
|
v_lang: list[str],
|
||||||
a_lang: list[str],
|
a_lang: list[str],
|
||||||
s_lang: list[str],
|
s_lang: list[str],
|
||||||
|
require_subs: list[str],
|
||||||
forced_subs: bool,
|
forced_subs: bool,
|
||||||
sub_format: Optional[Subtitle.Codec],
|
sub_format: Optional[Subtitle.Codec],
|
||||||
video_only: bool,
|
video_only: bool,
|
||||||
@@ -447,6 +486,7 @@ class dl:
|
|||||||
no_source: bool,
|
no_source: bool,
|
||||||
workers: Optional[int],
|
workers: Optional[int],
|
||||||
downloads: int,
|
downloads: int,
|
||||||
|
best_available: bool,
|
||||||
*_: Any,
|
*_: Any,
|
||||||
**__: Any,
|
**__: Any,
|
||||||
) -> None:
|
) -> None:
|
||||||
@@ -454,6 +494,10 @@ class dl:
|
|||||||
self.search_source = None
|
self.search_source = None
|
||||||
start_time = time.time()
|
start_time = time.time()
|
||||||
|
|
||||||
|
if require_subs and s_lang != ["all"]:
|
||||||
|
self.log.error("--require-subs and --s-lang cannot be used together")
|
||||||
|
sys.exit(1)
|
||||||
|
|
||||||
# Check if dovi_tool is available when hybrid mode is requested
|
# Check if dovi_tool is available when hybrid mode is requested
|
||||||
if any(r == Video.Range.HYBRID for r in range_):
|
if any(r == Video.Range.HYBRID for r in range_):
|
||||||
from unshackle.core.binaries import DoviTool
|
from unshackle.core.binaries import DoviTool
|
||||||
@@ -688,8 +732,14 @@ class dl:
|
|||||||
res_list = ", ".join([f"{x}p" for x in missing_resolutions[:-1]]) + " or "
|
res_list = ", ".join([f"{x}p" for x in missing_resolutions[:-1]]) + " or "
|
||||||
res_list = f"{res_list}{missing_resolutions[-1]}p"
|
res_list = f"{res_list}{missing_resolutions[-1]}p"
|
||||||
plural = "s" if len(missing_resolutions) > 1 else ""
|
plural = "s" if len(missing_resolutions) > 1 else ""
|
||||||
self.log.error(f"There's no {res_list} Video Track{plural}...")
|
|
||||||
sys.exit(1)
|
if best_available:
|
||||||
|
self.log.warning(
|
||||||
|
f"There's no {res_list} Video Track{plural}, continuing with available qualities..."
|
||||||
|
)
|
||||||
|
else:
|
||||||
|
self.log.error(f"There's no {res_list} Video Track{plural}...")
|
||||||
|
sys.exit(1)
|
||||||
|
|
||||||
# choose best track by range and quality
|
# choose best track by range and quality
|
||||||
if any(r == Video.Range.HYBRID for r in range_):
|
if any(r == Video.Range.HYBRID for r in range_):
|
||||||
@@ -725,7 +775,21 @@ class dl:
|
|||||||
title.tracks.videos = selected_videos
|
title.tracks.videos = selected_videos
|
||||||
|
|
||||||
# filter subtitle tracks
|
# filter subtitle tracks
|
||||||
if s_lang and "all" not in s_lang:
|
if require_subs:
|
||||||
|
missing_langs = [
|
||||||
|
lang
|
||||||
|
for lang in require_subs
|
||||||
|
if not any(is_close_match(lang, [sub.language]) for sub in title.tracks.subtitles)
|
||||||
|
]
|
||||||
|
|
||||||
|
if missing_langs:
|
||||||
|
self.log.error(f"Required subtitle language(s) not found: {', '.join(missing_langs)}")
|
||||||
|
sys.exit(1)
|
||||||
|
|
||||||
|
self.log.info(
|
||||||
|
f"Required languages found ({', '.join(require_subs)}), downloading all available subtitles"
|
||||||
|
)
|
||||||
|
elif s_lang and "all" not in s_lang:
|
||||||
missing_langs = [
|
missing_langs = [
|
||||||
lang_
|
lang_
|
||||||
for lang_ in s_lang
|
for lang_ in s_lang
|
||||||
@@ -847,9 +911,40 @@ class dl:
|
|||||||
|
|
||||||
selected_tracks, tracks_progress_callables = title.tracks.tree(add_progress=True)
|
selected_tracks, tracks_progress_callables = title.tracks.tree(add_progress=True)
|
||||||
|
|
||||||
|
for track in title.tracks:
|
||||||
|
if hasattr(track, "needs_drm_loading") and track.needs_drm_loading:
|
||||||
|
track.load_drm_if_needed(service)
|
||||||
|
|
||||||
download_table = Table.grid()
|
download_table = Table.grid()
|
||||||
download_table.add_row(selected_tracks)
|
download_table.add_row(selected_tracks)
|
||||||
|
|
||||||
|
video_tracks = title.tracks.videos
|
||||||
|
if video_tracks:
|
||||||
|
highest_quality = max((track.height for track in video_tracks if track.height), default=0)
|
||||||
|
if highest_quality > 0:
|
||||||
|
if isinstance(self.cdm, (WidevineCdm, DecryptLabsRemoteCDM)) and not (
|
||||||
|
isinstance(self.cdm, DecryptLabsRemoteCDM) and self.cdm.is_playready
|
||||||
|
):
|
||||||
|
quality_based_cdm = self.get_cdm(
|
||||||
|
self.service, self.profile, drm="widevine", quality=highest_quality
|
||||||
|
)
|
||||||
|
if quality_based_cdm and quality_based_cdm != self.cdm:
|
||||||
|
self.log.debug(
|
||||||
|
f"Pre-selecting Widevine CDM based on highest quality {highest_quality}p across all video tracks"
|
||||||
|
)
|
||||||
|
self.cdm = quality_based_cdm
|
||||||
|
elif isinstance(self.cdm, (PlayReadyCdm, DecryptLabsRemoteCDM)) and (
|
||||||
|
isinstance(self.cdm, DecryptLabsRemoteCDM) and self.cdm.is_playready
|
||||||
|
):
|
||||||
|
quality_based_cdm = self.get_cdm(
|
||||||
|
self.service, self.profile, drm="playready", quality=highest_quality
|
||||||
|
)
|
||||||
|
if quality_based_cdm and quality_based_cdm != self.cdm:
|
||||||
|
self.log.debug(
|
||||||
|
f"Pre-selecting PlayReady CDM based on highest quality {highest_quality}p across all video tracks"
|
||||||
|
)
|
||||||
|
self.cdm = quality_based_cdm
|
||||||
|
|
||||||
dl_start_time = time.time()
|
dl_start_time = time.time()
|
||||||
|
|
||||||
if skip_dl:
|
if skip_dl:
|
||||||
@@ -874,7 +969,12 @@ class dl:
|
|||||||
),
|
),
|
||||||
licence=partial(
|
licence=partial(
|
||||||
service.get_playready_license
|
service.get_playready_license
|
||||||
if isinstance(self.cdm, PlayReadyCdm)
|
if (
|
||||||
|
isinstance(self.cdm, PlayReadyCdm)
|
||||||
|
or (
|
||||||
|
isinstance(self.cdm, DecryptLabsRemoteCDM) and self.cdm.is_playready
|
||||||
|
)
|
||||||
|
)
|
||||||
and hasattr(service, "get_playready_license")
|
and hasattr(service, "get_playready_license")
|
||||||
else service.get_widevine_license,
|
else service.get_widevine_license,
|
||||||
title=title,
|
title=title,
|
||||||
@@ -1127,8 +1227,13 @@ class dl:
|
|||||||
with Live(Padding(progress, (0, 5, 1, 5)), console=console):
|
with Live(Padding(progress, (0, 5, 1, 5)), console=console):
|
||||||
for task_id, task_tracks in multiplex_tasks:
|
for task_id, task_tracks in multiplex_tasks:
|
||||||
progress.start_task(task_id) # TODO: Needed?
|
progress.start_task(task_id) # TODO: Needed?
|
||||||
|
audio_expected = not video_only and not no_audio
|
||||||
muxed_path, return_code, errors = task_tracks.mux(
|
muxed_path, return_code, errors = task_tracks.mux(
|
||||||
str(title), progress=partial(progress.update, task_id=task_id), delete=False
|
str(title),
|
||||||
|
progress=partial(progress.update, task_id=task_id),
|
||||||
|
delete=False,
|
||||||
|
audio_expected=audio_expected,
|
||||||
|
title_language=title.language,
|
||||||
)
|
)
|
||||||
muxed_paths.append(muxed_path)
|
muxed_paths.append(muxed_path)
|
||||||
if return_code >= 2:
|
if return_code >= 2:
|
||||||
@@ -1201,21 +1306,43 @@ class dl:
|
|||||||
if not drm:
|
if not drm:
|
||||||
return
|
return
|
||||||
|
|
||||||
if isinstance(drm, Widevine) and not isinstance(self.cdm, WidevineCdm):
|
if isinstance(track, Video) and track.height:
|
||||||
self.cdm = self.get_cdm(self.service, self.profile, drm="widevine")
|
pass
|
||||||
elif isinstance(drm, PlayReady) and not isinstance(self.cdm, PlayReadyCdm):
|
|
||||||
self.cdm = self.get_cdm(self.service, self.profile, drm="playready")
|
if isinstance(drm, Widevine):
|
||||||
|
if not isinstance(self.cdm, (WidevineCdm, DecryptLabsRemoteCDM)) or (
|
||||||
|
isinstance(self.cdm, DecryptLabsRemoteCDM) and self.cdm.is_playready
|
||||||
|
):
|
||||||
|
widevine_cdm = self.get_cdm(self.service, self.profile, drm="widevine")
|
||||||
|
if widevine_cdm:
|
||||||
|
self.log.info("Switching to Widevine CDM for Widevine content")
|
||||||
|
self.cdm = widevine_cdm
|
||||||
|
|
||||||
|
elif isinstance(drm, PlayReady):
|
||||||
|
if not isinstance(self.cdm, (PlayReadyCdm, DecryptLabsRemoteCDM)) or (
|
||||||
|
isinstance(self.cdm, DecryptLabsRemoteCDM) and not self.cdm.is_playready
|
||||||
|
):
|
||||||
|
playready_cdm = self.get_cdm(self.service, self.profile, drm="playready")
|
||||||
|
if playready_cdm:
|
||||||
|
self.log.info("Switching to PlayReady CDM for PlayReady content")
|
||||||
|
self.cdm = playready_cdm
|
||||||
|
|
||||||
if isinstance(drm, Widevine):
|
if isinstance(drm, Widevine):
|
||||||
with self.DRM_TABLE_LOCK:
|
with self.DRM_TABLE_LOCK:
|
||||||
cek_tree = Tree(Text.assemble(("Widevine", "cyan"), (f"({drm.pssh.dumps()})", "text"), overflow="fold"))
|
pssh_display = self._truncate_pssh_for_display(drm.pssh.dumps(), "Widevine")
|
||||||
|
cek_tree = Tree(Text.assemble(("Widevine", "cyan"), (f"({pssh_display})", "text"), overflow="fold"))
|
||||||
pre_existing_tree = next(
|
pre_existing_tree = next(
|
||||||
(x for x in table.columns[0].cells if isinstance(x, Tree) and x.label == cek_tree.label), None
|
(x for x in table.columns[0].cells if isinstance(x, Tree) and x.label == cek_tree.label), None
|
||||||
)
|
)
|
||||||
if pre_existing_tree:
|
if pre_existing_tree:
|
||||||
cek_tree = pre_existing_tree
|
cek_tree = pre_existing_tree
|
||||||
|
|
||||||
for kid in drm.kids:
|
need_license = False
|
||||||
|
all_kids = list(drm.kids)
|
||||||
|
if track_kid and track_kid not in all_kids:
|
||||||
|
all_kids.append(track_kid)
|
||||||
|
|
||||||
|
for kid in all_kids:
|
||||||
if kid in drm.content_keys:
|
if kid in drm.content_keys:
|
||||||
continue
|
continue
|
||||||
|
|
||||||
@@ -1235,46 +1362,51 @@ class dl:
|
|||||||
if not pre_existing_tree:
|
if not pre_existing_tree:
|
||||||
table.add_row(cek_tree)
|
table.add_row(cek_tree)
|
||||||
raise Widevine.Exceptions.CEKNotFound(msg)
|
raise Widevine.Exceptions.CEKNotFound(msg)
|
||||||
|
else:
|
||||||
|
need_license = True
|
||||||
|
|
||||||
if kid not in drm.content_keys and not vaults_only:
|
if kid not in drm.content_keys and cdm_only:
|
||||||
from_vaults = drm.content_keys.copy()
|
need_license = True
|
||||||
|
|
||||||
try:
|
if need_license and not vaults_only:
|
||||||
if self.service == "NF":
|
from_vaults = drm.content_keys.copy()
|
||||||
drm.get_NF_content_keys(cdm=self.cdm, licence=licence, certificate=certificate)
|
|
||||||
else:
|
|
||||||
drm.get_content_keys(cdm=self.cdm, licence=licence, certificate=certificate)
|
|
||||||
except Exception as e:
|
|
||||||
if isinstance(e, (Widevine.Exceptions.EmptyLicense, Widevine.Exceptions.CEKNotFound)):
|
|
||||||
msg = str(e)
|
|
||||||
else:
|
|
||||||
msg = f"An exception occurred in the Service's license function: {e}"
|
|
||||||
cek_tree.add(f"[logging.level.error]{msg}")
|
|
||||||
if not pre_existing_tree:
|
|
||||||
table.add_row(cek_tree)
|
|
||||||
raise e
|
|
||||||
|
|
||||||
for kid_, key in drm.content_keys.items():
|
try:
|
||||||
if key == "0" * 32:
|
if self.service == "NF":
|
||||||
key = f"[red]{key}[/]"
|
drm.get_NF_content_keys(cdm=self.cdm, licence=licence, certificate=certificate)
|
||||||
label = f"[text2]{kid_.hex}:{key}{is_track_kid}"
|
else:
|
||||||
if not any(f"{kid_.hex}:{key}" in x.label for x in cek_tree.children):
|
drm.get_content_keys(cdm=self.cdm, licence=licence, certificate=certificate)
|
||||||
cek_tree.add(label)
|
except Exception as e:
|
||||||
|
if isinstance(e, (Widevine.Exceptions.EmptyLicense, Widevine.Exceptions.CEKNotFound)):
|
||||||
|
msg = str(e)
|
||||||
|
else:
|
||||||
|
msg = f"An exception occurred in the Service's license function: {e}"
|
||||||
|
cek_tree.add(f"[logging.level.error]{msg}")
|
||||||
|
if not pre_existing_tree:
|
||||||
|
table.add_row(cek_tree)
|
||||||
|
raise e
|
||||||
|
|
||||||
drm.content_keys = {
|
for kid_, key in drm.content_keys.items():
|
||||||
kid_: key for kid_, key in drm.content_keys.items() if key and key.count("0") != len(key)
|
if key == "0" * 32:
|
||||||
}
|
key = f"[red]{key}[/]"
|
||||||
|
is_track_kid_marker = ["", "*"][kid_ == track_kid]
|
||||||
|
label = f"[text2]{kid_.hex}:{key}{is_track_kid_marker}"
|
||||||
|
if not any(f"{kid_.hex}:{key}" in x.label for x in cek_tree.children):
|
||||||
|
cek_tree.add(label)
|
||||||
|
|
||||||
# The CDM keys may have returned blank content keys for KIDs we got from vaults.
|
drm.content_keys = {
|
||||||
# So we re-add the keys from vaults earlier overwriting blanks or removed KIDs data.
|
kid_: key for kid_, key in drm.content_keys.items() if key and key.count("0") != len(key)
|
||||||
drm.content_keys.update(from_vaults)
|
}
|
||||||
|
|
||||||
successful_caches = self.vaults.add_keys(drm.content_keys)
|
# The CDM keys may have returned blank content keys for KIDs we got from vaults.
|
||||||
self.log.info(
|
# So we re-add the keys from vaults earlier overwriting blanks or removed KIDs data.
|
||||||
f"Cached {len(drm.content_keys)} Key{'' if len(drm.content_keys) == 1 else 's'} to "
|
drm.content_keys.update(from_vaults)
|
||||||
f"{successful_caches}/{len(self.vaults)} Vaults"
|
|
||||||
)
|
successful_caches = self.vaults.add_keys(drm.content_keys)
|
||||||
break # licensing twice will be unnecessary
|
self.log.info(
|
||||||
|
f"Cached {len(drm.content_keys)} Key{'' if len(drm.content_keys) == 1 else 's'} to "
|
||||||
|
f"{successful_caches}/{len(self.vaults)} Vaults"
|
||||||
|
)
|
||||||
|
|
||||||
if track_kid and track_kid not in drm.content_keys:
|
if track_kid and track_kid not in drm.content_keys:
|
||||||
msg = f"No Content Key for KID {track_kid.hex} was returned in the License"
|
msg = f"No Content Key for KID {track_kid.hex} was returned in the License"
|
||||||
@@ -1300,10 +1432,11 @@ class dl:
|
|||||||
|
|
||||||
elif isinstance(drm, PlayReady):
|
elif isinstance(drm, PlayReady):
|
||||||
with self.DRM_TABLE_LOCK:
|
with self.DRM_TABLE_LOCK:
|
||||||
|
pssh_display = self._truncate_pssh_for_display(drm.pssh_b64 or "", "PlayReady")
|
||||||
cek_tree = Tree(
|
cek_tree = Tree(
|
||||||
Text.assemble(
|
Text.assemble(
|
||||||
("PlayReady", "cyan"),
|
("PlayReady", "cyan"),
|
||||||
(f"({drm.pssh_b64 or ''})", "text"),
|
(f"({pssh_display})", "text"),
|
||||||
overflow="fold",
|
overflow="fold",
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
@@ -1313,7 +1446,12 @@ class dl:
|
|||||||
if pre_existing_tree:
|
if pre_existing_tree:
|
||||||
cek_tree = pre_existing_tree
|
cek_tree = pre_existing_tree
|
||||||
|
|
||||||
for kid in drm.kids:
|
need_license = False
|
||||||
|
all_kids = list(drm.kids)
|
||||||
|
if track_kid and track_kid not in all_kids:
|
||||||
|
all_kids.append(track_kid)
|
||||||
|
|
||||||
|
for kid in all_kids:
|
||||||
if kid in drm.content_keys:
|
if kid in drm.content_keys:
|
||||||
continue
|
continue
|
||||||
|
|
||||||
@@ -1333,35 +1471,40 @@ class dl:
|
|||||||
if not pre_existing_tree:
|
if not pre_existing_tree:
|
||||||
table.add_row(cek_tree)
|
table.add_row(cek_tree)
|
||||||
raise PlayReady.Exceptions.CEKNotFound(msg)
|
raise PlayReady.Exceptions.CEKNotFound(msg)
|
||||||
|
else:
|
||||||
|
need_license = True
|
||||||
|
|
||||||
if kid not in drm.content_keys and not vaults_only:
|
if kid not in drm.content_keys and cdm_only:
|
||||||
from_vaults = drm.content_keys.copy()
|
need_license = True
|
||||||
|
|
||||||
try:
|
if need_license and not vaults_only:
|
||||||
drm.get_content_keys(cdm=self.cdm, licence=licence, certificate=certificate)
|
from_vaults = drm.content_keys.copy()
|
||||||
except Exception as e:
|
|
||||||
if isinstance(e, (PlayReady.Exceptions.EmptyLicense, PlayReady.Exceptions.CEKNotFound)):
|
|
||||||
msg = str(e)
|
|
||||||
else:
|
|
||||||
msg = f"An exception occurred in the Service's license function: {e}"
|
|
||||||
cek_tree.add(f"[logging.level.error]{msg}")
|
|
||||||
if not pre_existing_tree:
|
|
||||||
table.add_row(cek_tree)
|
|
||||||
raise e
|
|
||||||
|
|
||||||
for kid_, key in drm.content_keys.items():
|
try:
|
||||||
label = f"[text2]{kid_.hex}:{key}{is_track_kid}"
|
drm.get_content_keys(cdm=self.cdm, licence=licence, certificate=certificate)
|
||||||
if not any(f"{kid_.hex}:{key}" in x.label for x in cek_tree.children):
|
except Exception as e:
|
||||||
cek_tree.add(label)
|
if isinstance(e, (PlayReady.Exceptions.EmptyLicense, PlayReady.Exceptions.CEKNotFound)):
|
||||||
|
msg = str(e)
|
||||||
|
else:
|
||||||
|
msg = f"An exception occurred in the Service's license function: {e}"
|
||||||
|
cek_tree.add(f"[logging.level.error]{msg}")
|
||||||
|
if not pre_existing_tree:
|
||||||
|
table.add_row(cek_tree)
|
||||||
|
raise e
|
||||||
|
|
||||||
drm.content_keys.update(from_vaults)
|
for kid_, key in drm.content_keys.items():
|
||||||
|
is_track_kid_marker = ["", "*"][kid_ == track_kid]
|
||||||
|
label = f"[text2]{kid_.hex}:{key}{is_track_kid_marker}"
|
||||||
|
if not any(f"{kid_.hex}:{key}" in x.label for x in cek_tree.children):
|
||||||
|
cek_tree.add(label)
|
||||||
|
|
||||||
successful_caches = self.vaults.add_keys(drm.content_keys)
|
drm.content_keys.update(from_vaults)
|
||||||
self.log.info(
|
|
||||||
f"Cached {len(drm.content_keys)} Key{'' if len(drm.content_keys) == 1 else 's'} to "
|
successful_caches = self.vaults.add_keys(drm.content_keys)
|
||||||
f"{successful_caches}/{len(self.vaults)} Vaults"
|
self.log.info(
|
||||||
)
|
f"Cached {len(drm.content_keys)} Key{'' if len(drm.content_keys) == 1 else 's'} to "
|
||||||
break
|
f"{successful_caches}/{len(self.vaults)} Vaults"
|
||||||
|
)
|
||||||
|
|
||||||
if track_kid and track_kid not in drm.content_keys:
|
if track_kid and track_kid not in drm.content_keys:
|
||||||
msg = f"No Content Key for KID {track_kid.hex} was returned in the License"
|
msg = f"No Content Key for KID {track_kid.hex} was returned in the License"
|
||||||
@@ -1421,6 +1564,9 @@ class dl:
|
|||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def save_cookies(path: Path, cookies: CookieJar):
|
def save_cookies(path: Path, cookies: CookieJar):
|
||||||
|
if hasattr(cookies, 'jar'):
|
||||||
|
cookies = cookies.jar
|
||||||
|
|
||||||
cookie_jar = MozillaCookieJar(path)
|
cookie_jar = MozillaCookieJar(path)
|
||||||
cookie_jar.load()
|
cookie_jar.load()
|
||||||
for cookie in cookies:
|
for cookie in cookies:
|
||||||
@@ -1442,14 +1588,16 @@ class dl:
|
|||||||
return Credential(*credentials)
|
return Credential(*credentials)
|
||||||
return Credential.loads(credentials) # type: ignore
|
return Credential.loads(credentials) # type: ignore
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def get_cdm(
|
def get_cdm(
|
||||||
|
self,
|
||||||
service: str,
|
service: str,
|
||||||
profile: Optional[str] = None,
|
profile: Optional[str] = None,
|
||||||
drm: Optional[str] = None,
|
drm: Optional[str] = None,
|
||||||
|
quality: Optional[int] = None,
|
||||||
) -> Optional[object]:
|
) -> Optional[object]:
|
||||||
"""
|
"""
|
||||||
Get CDM for a specified service (either Local or Remote CDM).
|
Get CDM for a specified service (either Local or Remote CDM).
|
||||||
|
Now supports quality-based selection when quality is provided.
|
||||||
Raises a ValueError if there's a problem getting a CDM.
|
Raises a ValueError if there's a problem getting a CDM.
|
||||||
"""
|
"""
|
||||||
cdm_name = config.cdm.get(service) or config.cdm.get("default")
|
cdm_name = config.cdm.get(service) or config.cdm.get("default")
|
||||||
@@ -1457,30 +1605,106 @@ class dl:
|
|||||||
return None
|
return None
|
||||||
|
|
||||||
if isinstance(cdm_name, dict):
|
if isinstance(cdm_name, dict):
|
||||||
lower_keys = {k.lower(): v for k, v in cdm_name.items()}
|
if quality:
|
||||||
if {"widevine", "playready"} & lower_keys.keys():
|
quality_match = None
|
||||||
drm_key = None
|
quality_keys = []
|
||||||
if drm:
|
|
||||||
drm_key = {
|
|
||||||
"wv": "widevine",
|
|
||||||
"widevine": "widevine",
|
|
||||||
"pr": "playready",
|
|
||||||
"playready": "playready",
|
|
||||||
}.get(drm.lower())
|
|
||||||
cdm_name = lower_keys.get(drm_key or "widevine") or lower_keys.get("playready")
|
|
||||||
else:
|
|
||||||
if not profile:
|
|
||||||
return None
|
|
||||||
cdm_name = cdm_name.get(profile) or config.cdm.get("default")
|
|
||||||
if not cdm_name:
|
|
||||||
return None
|
|
||||||
|
|
||||||
cdm_api = next(iter(x for x in config.remote_cdm if x["name"] == cdm_name), None)
|
for key in cdm_name.keys():
|
||||||
|
if (
|
||||||
|
isinstance(key, str)
|
||||||
|
and any(op in key for op in [">=", ">", "<=", "<"])
|
||||||
|
or (isinstance(key, str) and key.isdigit())
|
||||||
|
):
|
||||||
|
quality_keys.append(key)
|
||||||
|
|
||||||
|
def sort_quality_key(key):
|
||||||
|
if key.isdigit():
|
||||||
|
return (0, int(key)) # Exact matches first
|
||||||
|
elif key.startswith(">="):
|
||||||
|
return (1, -int(key[2:])) # >= descending
|
||||||
|
elif key.startswith(">"):
|
||||||
|
return (1, -int(key[1:])) # > descending
|
||||||
|
elif key.startswith("<="):
|
||||||
|
return (2, int(key[2:])) # <= ascending
|
||||||
|
elif key.startswith("<"):
|
||||||
|
return (2, int(key[1:])) # < ascending
|
||||||
|
return (3, 0) # Other keys last
|
||||||
|
|
||||||
|
quality_keys.sort(key=sort_quality_key)
|
||||||
|
|
||||||
|
for key in quality_keys:
|
||||||
|
if key.isdigit() and quality == int(key):
|
||||||
|
quality_match = cdm_name[key]
|
||||||
|
self.log.debug(f"Selected CDM based on exact quality match {quality}p: {quality_match}")
|
||||||
|
break
|
||||||
|
elif key.startswith(">="):
|
||||||
|
threshold = int(key[2:])
|
||||||
|
if quality >= threshold:
|
||||||
|
quality_match = cdm_name[key]
|
||||||
|
self.log.debug(f"Selected CDM based on quality {quality}p >= {threshold}p: {quality_match}")
|
||||||
|
break
|
||||||
|
elif key.startswith(">"):
|
||||||
|
threshold = int(key[1:])
|
||||||
|
if quality > threshold:
|
||||||
|
quality_match = cdm_name[key]
|
||||||
|
self.log.debug(f"Selected CDM based on quality {quality}p > {threshold}p: {quality_match}")
|
||||||
|
break
|
||||||
|
elif key.startswith("<="):
|
||||||
|
threshold = int(key[2:])
|
||||||
|
if quality <= threshold:
|
||||||
|
quality_match = cdm_name[key]
|
||||||
|
self.log.debug(f"Selected CDM based on quality {quality}p <= {threshold}p: {quality_match}")
|
||||||
|
break
|
||||||
|
elif key.startswith("<"):
|
||||||
|
threshold = int(key[1:])
|
||||||
|
if quality < threshold:
|
||||||
|
quality_match = cdm_name[key]
|
||||||
|
self.log.debug(f"Selected CDM based on quality {quality}p < {threshold}p: {quality_match}")
|
||||||
|
break
|
||||||
|
|
||||||
|
if quality_match:
|
||||||
|
cdm_name = quality_match
|
||||||
|
|
||||||
|
if isinstance(cdm_name, dict):
|
||||||
|
lower_keys = {k.lower(): v for k, v in cdm_name.items()}
|
||||||
|
if {"widevine", "playready"} & lower_keys.keys():
|
||||||
|
drm_key = None
|
||||||
|
if drm:
|
||||||
|
drm_key = {
|
||||||
|
"wv": "widevine",
|
||||||
|
"widevine": "widevine",
|
||||||
|
"pr": "playready",
|
||||||
|
"playready": "playready",
|
||||||
|
}.get(drm.lower())
|
||||||
|
cdm_name = lower_keys.get(drm_key or "widevine") or lower_keys.get("playready")
|
||||||
|
else:
|
||||||
|
cdm_name = cdm_name.get(profile) or cdm_name.get("default") or config.cdm.get("default")
|
||||||
|
if not cdm_name:
|
||||||
|
return None
|
||||||
|
|
||||||
|
cdm_api = next(iter(x.copy() for x in config.remote_cdm if x["name"] == cdm_name), None)
|
||||||
if cdm_api:
|
if cdm_api:
|
||||||
is_decrypt_lab = True if cdm_api["type"] == "decrypt_labs" else False
|
is_decrypt_lab = True if cdm_api.get("type") == "decrypt_labs" else False
|
||||||
del cdm_api["name"]
|
if is_decrypt_lab:
|
||||||
del cdm_api["type"]
|
del cdm_api["name"]
|
||||||
return DecryptLabsRemoteCDM(service_name=service, **cdm_api) if is_decrypt_lab else RemoteCdm(**cdm_api)
|
del cdm_api["type"]
|
||||||
|
|
||||||
|
if "secret" not in cdm_api or not cdm_api["secret"]:
|
||||||
|
if config.decrypt_labs_api_key:
|
||||||
|
cdm_api["secret"] = config.decrypt_labs_api_key
|
||||||
|
else:
|
||||||
|
raise ValueError(
|
||||||
|
f"No secret provided for DecryptLabs CDM '{cdm_name}' and no global "
|
||||||
|
"decrypt_labs_api_key configured"
|
||||||
|
)
|
||||||
|
|
||||||
|
# All DecryptLabs CDMs use DecryptLabsRemoteCDM
|
||||||
|
return DecryptLabsRemoteCDM(service_name=service, vaults=self.vaults, **cdm_api)
|
||||||
|
else:
|
||||||
|
del cdm_api["name"]
|
||||||
|
if "type" in cdm_api:
|
||||||
|
del cdm_api["type"]
|
||||||
|
return RemoteCdm(**cdm_api)
|
||||||
|
|
||||||
prd_path = config.directories.prds / f"{cdm_name}.prd"
|
prd_path = config.directories.prds / f"{cdm_name}.prd"
|
||||||
if not prd_path.is_file():
|
if not prd_path.is_file():
|
||||||
|
|||||||
@@ -12,84 +12,113 @@ from unshackle.core.vault import Vault
|
|||||||
from unshackle.core.vaults import Vaults
|
from unshackle.core.vaults import Vaults
|
||||||
|
|
||||||
|
|
||||||
|
def _load_vaults(vault_names: list[str]) -> Vaults:
|
||||||
|
"""Load and validate vaults by name."""
|
||||||
|
vaults = Vaults()
|
||||||
|
for vault_name in vault_names:
|
||||||
|
vault_config = next((x for x in config.key_vaults if x["name"] == vault_name), None)
|
||||||
|
if not vault_config:
|
||||||
|
raise click.ClickException(f"Vault ({vault_name}) is not defined in the config.")
|
||||||
|
|
||||||
|
vault_type = vault_config["type"]
|
||||||
|
vault_args = vault_config.copy()
|
||||||
|
del vault_args["type"]
|
||||||
|
|
||||||
|
if not vaults.load(vault_type, **vault_args):
|
||||||
|
raise click.ClickException(f"Failed to load vault ({vault_name}).")
|
||||||
|
|
||||||
|
return vaults
|
||||||
|
|
||||||
|
|
||||||
|
def _process_service_keys(from_vault: Vault, service: str, log: logging.Logger) -> dict[str, str]:
|
||||||
|
"""Get and validate keys from a vault for a specific service."""
|
||||||
|
content_keys = list(from_vault.get_keys(service))
|
||||||
|
|
||||||
|
bad_keys = {kid: key for kid, key in content_keys if not key or key.count("0") == len(key)}
|
||||||
|
for kid, key in bad_keys.items():
|
||||||
|
log.warning(f"Skipping NULL key: {kid}:{key}")
|
||||||
|
|
||||||
|
return {kid: key for kid, key in content_keys if kid not in bad_keys}
|
||||||
|
|
||||||
|
|
||||||
|
def _copy_service_data(to_vault: Vault, from_vault: Vault, service: str, log: logging.Logger) -> int:
|
||||||
|
"""Copy data for a single service between vaults."""
|
||||||
|
content_keys = _process_service_keys(from_vault, service, log)
|
||||||
|
total_count = len(content_keys)
|
||||||
|
|
||||||
|
if total_count == 0:
|
||||||
|
log.info(f"{service}: No keys found in {from_vault}")
|
||||||
|
return 0
|
||||||
|
|
||||||
|
try:
|
||||||
|
added = to_vault.add_keys(service, content_keys)
|
||||||
|
except PermissionError:
|
||||||
|
log.warning(f"{service}: No permission to create table in {to_vault}, skipped")
|
||||||
|
return 0
|
||||||
|
|
||||||
|
existed = total_count - added
|
||||||
|
|
||||||
|
if added > 0 and existed > 0:
|
||||||
|
log.info(f"{service}: {added} added, {existed} skipped ({total_count} total)")
|
||||||
|
elif added > 0:
|
||||||
|
log.info(f"{service}: {added} added ({total_count} total)")
|
||||||
|
else:
|
||||||
|
log.info(f"{service}: {existed} skipped (all existed)")
|
||||||
|
|
||||||
|
return added
|
||||||
|
|
||||||
|
|
||||||
@click.group(short_help="Manage and configure Key Vaults.", context_settings=context_settings)
|
@click.group(short_help="Manage and configure Key Vaults.", context_settings=context_settings)
|
||||||
def kv() -> None:
|
def kv() -> None:
|
||||||
"""Manage and configure Key Vaults."""
|
"""Manage and configure Key Vaults."""
|
||||||
|
|
||||||
|
|
||||||
@kv.command()
|
@kv.command()
|
||||||
@click.argument("to_vault", type=str)
|
@click.argument("to_vault_name", type=str)
|
||||||
@click.argument("from_vaults", nargs=-1, type=click.UNPROCESSED)
|
@click.argument("from_vault_names", nargs=-1, type=click.UNPROCESSED)
|
||||||
@click.option("-s", "--service", type=str, default=None, help="Only copy data to and from a specific service.")
|
@click.option("-s", "--service", type=str, default=None, help="Only copy data to and from a specific service.")
|
||||||
def copy(to_vault: str, from_vaults: list[str], service: Optional[str] = None) -> None:
|
def copy(to_vault_name: str, from_vault_names: list[str], service: Optional[str] = None) -> None:
|
||||||
"""
|
"""
|
||||||
Copy data from multiple Key Vaults into a single Key Vault.
|
Copy data from multiple Key Vaults into a single Key Vault.
|
||||||
Rows with matching KIDs are skipped unless there's no KEY set.
|
Rows with matching KIDs are skipped unless there's no KEY set.
|
||||||
Existing data is not deleted or altered.
|
Existing data is not deleted or altered.
|
||||||
|
|
||||||
The `to_vault` argument is the key vault you wish to copy data to.
|
The `to_vault_name` argument is the key vault you wish to copy data to.
|
||||||
It should be the name of a Key Vault defined in the config.
|
It should be the name of a Key Vault defined in the config.
|
||||||
|
|
||||||
The `from_vaults` argument is the key vault(s) you wish to take
|
The `from_vault_names` argument is the key vault(s) you wish to take
|
||||||
data from. You may supply multiple key vaults.
|
data from. You may supply multiple key vaults.
|
||||||
"""
|
"""
|
||||||
if not from_vaults:
|
if not from_vault_names:
|
||||||
raise click.ClickException("No Vaults were specified to copy data from.")
|
raise click.ClickException("No Vaults were specified to copy data from.")
|
||||||
|
|
||||||
log = logging.getLogger("kv")
|
log = logging.getLogger("kv")
|
||||||
|
|
||||||
vaults = Vaults()
|
all_vault_names = [to_vault_name] + list(from_vault_names)
|
||||||
for vault_name in [to_vault] + list(from_vaults):
|
vaults = _load_vaults(all_vault_names)
|
||||||
vault = next((x for x in config.key_vaults if x["name"] == vault_name), None)
|
|
||||||
if not vault:
|
|
||||||
raise click.ClickException(f"Vault ({vault_name}) is not defined in the config.")
|
|
||||||
vault_type = vault["type"]
|
|
||||||
vault_args = vault.copy()
|
|
||||||
del vault_args["type"]
|
|
||||||
if not vaults.load(vault_type, **vault_args):
|
|
||||||
raise click.ClickException(f"Failed to load vault ({vault_name}).")
|
|
||||||
|
|
||||||
to_vault: Vault = vaults.vaults[0]
|
to_vault = vaults.vaults[0]
|
||||||
from_vaults: list[Vault] = vaults.vaults[1:]
|
from_vaults = vaults.vaults[1:]
|
||||||
|
|
||||||
|
vault_names = ", ".join([v.name for v in from_vaults])
|
||||||
|
log.info(f"Copying data from {vault_names} → {to_vault.name}")
|
||||||
|
|
||||||
log.info(f"Copying data from {', '.join([x.name for x in from_vaults])}, into {to_vault.name}")
|
|
||||||
if service:
|
if service:
|
||||||
service = Services.get_tag(service)
|
service = Services.get_tag(service)
|
||||||
log.info(f"Only copying data for service {service}")
|
log.info(f"Filtering by service: {service}")
|
||||||
|
|
||||||
total_added = 0
|
total_added = 0
|
||||||
for from_vault in from_vaults:
|
for from_vault in from_vaults:
|
||||||
if service:
|
services_to_copy = [service] if service else from_vault.get_services()
|
||||||
services = [service]
|
|
||||||
else:
|
|
||||||
services = from_vault.get_services()
|
|
||||||
|
|
||||||
for service_ in services:
|
|
||||||
log.info(f"Getting data from {from_vault} for {service_}")
|
|
||||||
content_keys = list(from_vault.get_keys(service_)) # important as it's a generator we iterate twice
|
|
||||||
|
|
||||||
bad_keys = {kid: key for kid, key in content_keys if not key or key.count("0") == len(key)}
|
|
||||||
|
|
||||||
for kid, key in bad_keys.items():
|
|
||||||
log.warning(f"Cannot add a NULL Content Key to a Vault, skipping: {kid}:{key}")
|
|
||||||
|
|
||||||
content_keys = {kid: key for kid, key in content_keys if kid not in bad_keys}
|
|
||||||
|
|
||||||
total_count = len(content_keys)
|
|
||||||
log.info(f"Adding {total_count} Content Keys to {to_vault} for {service_}")
|
|
||||||
|
|
||||||
try:
|
|
||||||
added = to_vault.add_keys(service_, content_keys)
|
|
||||||
except PermissionError:
|
|
||||||
log.warning(f" - No permission to create table ({service_}) in {to_vault}, skipping...")
|
|
||||||
continue
|
|
||||||
|
|
||||||
|
for service_tag in services_to_copy:
|
||||||
|
added = _copy_service_data(to_vault, from_vault, service_tag, log)
|
||||||
total_added += added
|
total_added += added
|
||||||
existed = total_count - added
|
|
||||||
|
|
||||||
log.info(f"{to_vault} ({service_}): {added} newly added, {existed} already existed (skipped)")
|
if total_added > 0:
|
||||||
|
log.info(f"Successfully added {total_added} new keys to {to_vault}")
|
||||||
log.info(f"{to_vault}: {total_added} total newly added")
|
else:
|
||||||
|
log.info("Copy completed - no new keys to add")
|
||||||
|
|
||||||
|
|
||||||
@kv.command()
|
@kv.command()
|
||||||
@@ -106,9 +135,9 @@ def sync(ctx: click.Context, vaults: list[str], service: Optional[str] = None) -
|
|||||||
if not len(vaults) > 1:
|
if not len(vaults) > 1:
|
||||||
raise click.ClickException("You must provide more than one Vault to sync.")
|
raise click.ClickException("You must provide more than one Vault to sync.")
|
||||||
|
|
||||||
ctx.invoke(copy, to_vault=vaults[0], from_vaults=vaults[1:], service=service)
|
ctx.invoke(copy, to_vault_name=vaults[0], from_vault_names=vaults[1:], service=service)
|
||||||
for i in range(1, len(vaults)):
|
for i in range(1, len(vaults)):
|
||||||
ctx.invoke(copy, to_vault=vaults[i], from_vaults=[vaults[i - 1]], service=service)
|
ctx.invoke(copy, to_vault_name=vaults[i], from_vault_names=[vaults[i - 1]], service=service)
|
||||||
|
|
||||||
|
|
||||||
@kv.command()
|
@kv.command()
|
||||||
@@ -135,15 +164,7 @@ def add(file: Path, service: str, vaults: list[str]) -> None:
|
|||||||
log = logging.getLogger("kv")
|
log = logging.getLogger("kv")
|
||||||
service = Services.get_tag(service)
|
service = Services.get_tag(service)
|
||||||
|
|
||||||
vaults_ = Vaults()
|
vaults_ = _load_vaults(list(vaults))
|
||||||
for vault_name in vaults:
|
|
||||||
vault = next((x for x in config.key_vaults if x["name"] == vault_name), None)
|
|
||||||
if not vault:
|
|
||||||
raise click.ClickException(f"Vault ({vault_name}) is not defined in the config.")
|
|
||||||
vault_type = vault["type"]
|
|
||||||
vault_args = vault.copy()
|
|
||||||
del vault_args["type"]
|
|
||||||
vaults_.load(vault_type, **vault_args)
|
|
||||||
|
|
||||||
data = file.read_text(encoding="utf8")
|
data = file.read_text(encoding="utf8")
|
||||||
kid_keys: dict[str, str] = {}
|
kid_keys: dict[str, str] = {}
|
||||||
@@ -173,15 +194,7 @@ def prepare(vaults: list[str]) -> None:
|
|||||||
"""Create Service Tables on Vaults if not yet created."""
|
"""Create Service Tables on Vaults if not yet created."""
|
||||||
log = logging.getLogger("kv")
|
log = logging.getLogger("kv")
|
||||||
|
|
||||||
vaults_ = Vaults()
|
vaults_ = _load_vaults(vaults)
|
||||||
for vault_name in vaults:
|
|
||||||
vault = next((x for x in config.key_vaults if x["name"] == vault_name), None)
|
|
||||||
if not vault:
|
|
||||||
raise click.ClickException(f"Vault ({vault_name}) is not defined in the config.")
|
|
||||||
vault_type = vault["type"]
|
|
||||||
vault_args = vault.copy()
|
|
||||||
del vault_args["type"]
|
|
||||||
vaults_.load(vault_type, **vault_args)
|
|
||||||
|
|
||||||
for vault in vaults_:
|
for vault in vaults_:
|
||||||
if hasattr(vault, "has_table") and hasattr(vault, "create_table"):
|
if hasattr(vault, "has_table") and hasattr(vault, "create_table"):
|
||||||
|
|||||||
@@ -1 +1 @@
|
|||||||
__version__ = "1.4.3"
|
__version__ = "1.4.7"
|
||||||
|
|||||||
@@ -1,189 +1,747 @@
|
|||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
import base64
|
import base64
|
||||||
import secrets
|
import secrets
|
||||||
from typing import Optional, Type, Union
|
from typing import Any, Dict, List, Optional, Union
|
||||||
from uuid import UUID
|
from uuid import UUID
|
||||||
|
|
||||||
import requests
|
import requests
|
||||||
from pywidevine import PSSH, Device, DeviceTypes, Key, RemoteCdm
|
from pywidevine.cdm import Cdm as WidevineCdm
|
||||||
from pywidevine.license_protocol_pb2 import SignedDrmCertificate, SignedMessage
|
from pywidevine.device import DeviceTypes
|
||||||
|
from requests import Session
|
||||||
|
|
||||||
# Copyright 2024 by DevYukine.
|
from unshackle.core import __version__
|
||||||
|
from unshackle.core.vaults import Vaults
|
||||||
|
|
||||||
|
|
||||||
class DecryptLabsRemoteCDM(RemoteCdm):
|
class MockCertificateChain:
|
||||||
|
"""Mock certificate chain for PlayReady compatibility."""
|
||||||
|
|
||||||
|
def __init__(self, name: str):
|
||||||
|
self._name = name
|
||||||
|
|
||||||
|
def get_name(self) -> str:
|
||||||
|
return self._name
|
||||||
|
|
||||||
|
|
||||||
|
class Key:
|
||||||
|
"""Key object compatible with pywidevine."""
|
||||||
|
|
||||||
|
def __init__(self, kid: str, key: str, type_: str = "CONTENT"):
|
||||||
|
if isinstance(kid, str):
|
||||||
|
clean_kid = kid.replace("-", "")
|
||||||
|
if len(clean_kid) == 32:
|
||||||
|
self.kid = UUID(hex=clean_kid)
|
||||||
|
else:
|
||||||
|
self.kid = UUID(hex=clean_kid.ljust(32, "0"))
|
||||||
|
else:
|
||||||
|
self.kid = kid
|
||||||
|
|
||||||
|
if isinstance(key, str):
|
||||||
|
self.key = bytes.fromhex(key)
|
||||||
|
else:
|
||||||
|
self.key = key
|
||||||
|
|
||||||
|
self.type = type_
|
||||||
|
|
||||||
|
|
||||||
|
class DecryptLabsRemoteCDMExceptions:
|
||||||
|
"""Exception classes for compatibility with pywidevine CDM."""
|
||||||
|
|
||||||
|
class InvalidSession(Exception):
|
||||||
|
"""Raised when session ID is invalid."""
|
||||||
|
|
||||||
|
class TooManySessions(Exception):
|
||||||
|
"""Raised when session limit is reached."""
|
||||||
|
|
||||||
|
class InvalidInitData(Exception):
|
||||||
|
"""Raised when PSSH/init data is invalid."""
|
||||||
|
|
||||||
|
class InvalidLicenseType(Exception):
|
||||||
|
"""Raised when license type is invalid."""
|
||||||
|
|
||||||
|
class InvalidLicenseMessage(Exception):
|
||||||
|
"""Raised when license message is invalid."""
|
||||||
|
|
||||||
|
class InvalidContext(Exception):
|
||||||
|
"""Raised when session has no context data."""
|
||||||
|
|
||||||
|
class SignatureMismatch(Exception):
|
||||||
|
"""Raised when signature verification fails."""
|
||||||
|
|
||||||
|
|
||||||
|
class DecryptLabsRemoteCDM:
|
||||||
|
"""
|
||||||
|
Decrypt Labs Remote CDM implementation with intelligent caching system.
|
||||||
|
|
||||||
|
This class provides a drop-in replacement for pywidevine's local CDM using
|
||||||
|
Decrypt Labs' KeyXtractor API service, enhanced with smart caching logic
|
||||||
|
that minimizes unnecessary license requests.
|
||||||
|
|
||||||
|
Key Features:
|
||||||
|
- Compatible with both Widevine and PlayReady DRM schemes
|
||||||
|
- Intelligent caching that compares required vs. available keys
|
||||||
|
- Optimized caching for L1/L2 devices (leverages API auto-optimization)
|
||||||
|
- Automatic key combination for mixed cache/license scenarios
|
||||||
|
- Seamless fallback to license requests when keys are missing
|
||||||
|
|
||||||
|
Intelligent Caching System:
|
||||||
|
1. DRM classes (PlayReady/Widevine) provide required KIDs via set_required_kids()
|
||||||
|
2. get_license_challenge() first checks for cached keys
|
||||||
|
3. For L1/L2 devices, always attempts cached keys first (API optimized)
|
||||||
|
4. If cached keys satisfy requirements, returns empty challenge (no license needed)
|
||||||
|
5. If keys are missing, makes targeted license request for remaining keys
|
||||||
|
6. parse_license() combines cached and license keys intelligently
|
||||||
|
"""
|
||||||
|
|
||||||
|
service_certificate_challenge = b"\x08\x04"
|
||||||
|
|
||||||
def __init__(
|
def __init__(
|
||||||
self,
|
self,
|
||||||
device_type: Union[DeviceTypes, str],
|
|
||||||
system_id: int,
|
|
||||||
security_level: int,
|
|
||||||
host: str,
|
|
||||||
secret: str,
|
secret: str,
|
||||||
device_name: str,
|
host: str = "https://keyxtractor.decryptlabs.com",
|
||||||
service_name: str,
|
device_name: str = "ChromeCDM",
|
||||||
|
service_name: Optional[str] = None,
|
||||||
|
vaults: Optional[Vaults] = None,
|
||||||
|
device_type: Optional[str] = None,
|
||||||
|
system_id: Optional[int] = None,
|
||||||
|
security_level: Optional[int] = None,
|
||||||
|
**kwargs,
|
||||||
):
|
):
|
||||||
self.response_counter = 0
|
"""
|
||||||
self.pssh = None
|
Initialize Decrypt Labs Remote CDM for Widevine and PlayReady schemes.
|
||||||
self.api_session_ids = {}
|
|
||||||
self.license_request = None
|
Args:
|
||||||
self.service_name = service_name
|
secret: Decrypt Labs API key (matches config format)
|
||||||
|
host: Decrypt Labs API host URL (matches config format)
|
||||||
|
device_name: DRM scheme (ChromeCDM, L1, L2 for Widevine; SL2, SL3 for PlayReady)
|
||||||
|
service_name: Service name for key caching and vault operations
|
||||||
|
vaults: Vaults instance for local key caching
|
||||||
|
device_type: Device type (CHROME, ANDROID, PLAYREADY) - for compatibility
|
||||||
|
system_id: System ID - for compatibility
|
||||||
|
security_level: Security level - for compatibility
|
||||||
|
"""
|
||||||
|
_ = kwargs
|
||||||
|
|
||||||
|
self.secret = secret
|
||||||
|
self.host = host.rstrip("/")
|
||||||
self.device_name = device_name
|
self.device_name = device_name
|
||||||
self.keys = {}
|
self.service_name = service_name or ""
|
||||||
self.scheme = "L1" if device_name == "L1" else "widevine"
|
self.vaults = vaults
|
||||||
try:
|
self.uch = self.host != "https://keyxtractor.decryptlabs.com"
|
||||||
super().__init__(device_type, system_id, security_level, host, secret, device_name)
|
|
||||||
except Exception:
|
|
||||||
pass
|
|
||||||
self.req_session = requests.Session()
|
|
||||||
self.req_session.headers.update({"decrypt-labs-api-key": secret})
|
|
||||||
|
|
||||||
@classmethod
|
self._device_type_str = device_type
|
||||||
def from_device(cls, device: Device) -> Type["DecryptLabsRemoteCDM"]:
|
if device_type:
|
||||||
raise NotImplementedError("You cannot load a DecryptLabsRemoteCDM from a local Device file.")
|
self.device_type = self._get_device_type_enum(device_type)
|
||||||
|
|
||||||
def open(self) -> bytes:
|
self._is_playready = (device_type and device_type.upper() == "PLAYREADY") or (device_name in ["SL2", "SL3"])
|
||||||
# We stub this method to return a random session ID for now, later we save the api session id and resolve by our random generated one.
|
|
||||||
return bytes.fromhex(secrets.token_hex(16))
|
|
||||||
|
|
||||||
def close(self, session_id: bytes) -> None:
|
if self._is_playready:
|
||||||
# We stub this method to do nothing.
|
self.system_id = system_id or 0
|
||||||
pass
|
self.security_level = security_level or (2000 if device_name == "SL2" else 3000)
|
||||||
|
else:
|
||||||
|
self.system_id = system_id or 26830
|
||||||
|
self.security_level = security_level or 3
|
||||||
|
|
||||||
def set_service_certificate(self, session_id: bytes, certificate: Optional[Union[bytes, str]]) -> str:
|
self._sessions: Dict[bytes, Dict[str, Any]] = {}
|
||||||
if isinstance(certificate, bytes):
|
self._pssh_b64 = None
|
||||||
certificate = base64.b64encode(certificate).decode()
|
self._required_kids: Optional[List[str]] = None
|
||||||
|
self._http_session = Session()
|
||||||
# certificate needs to be base64 to be sent off to the API.
|
self._http_session.headers.update(
|
||||||
# it needs to intentionally be kept as base64 encoded SignedMessage.
|
{
|
||||||
|
"decrypt-labs-api-key": self.secret,
|
||||||
self.req_session.signed_device_certificate = certificate
|
"Content-Type": "application/json",
|
||||||
self.req_session.privacy_mode = True
|
"User-Agent": f"unshackle-decrypt-labs-cdm/{__version__}",
|
||||||
|
}
|
||||||
return "success"
|
|
||||||
|
|
||||||
def get_service_certificate(self, session_id: bytes) -> Optional[SignedDrmCertificate]:
|
|
||||||
raise NotImplementedError("This method is not implemented in this CDM")
|
|
||||||
|
|
||||||
def get_license_challenge(
|
|
||||||
self, session_id: bytes, pssh: PSSH, license_type: str = "STREAMING", privacy_mode: bool = True
|
|
||||||
) -> bytes:
|
|
||||||
self.pssh = pssh
|
|
||||||
|
|
||||||
request_data = {
|
|
||||||
"init_data": self.pssh.dumps(),
|
|
||||||
"service_certificate": self.req_session.signed_device_certificate,
|
|
||||||
"scheme": self.scheme,
|
|
||||||
"service": self.service_name,
|
|
||||||
}
|
|
||||||
# Add required parameter for L1 scheme
|
|
||||||
if self.scheme == "L1":
|
|
||||||
request_data["get_cached_keys_if_exists"] = True
|
|
||||||
res = self.session(
|
|
||||||
self.host + "/get-request",
|
|
||||||
request_data,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
# Check if we got cached keys instead of a challenge
|
def _get_device_type_enum(self, device_type: str):
|
||||||
if res.get("message_type") == "cached-keys":
|
"""Convert device type string to enum for compatibility."""
|
||||||
# Store cached keys directly
|
device_type_upper = device_type.upper()
|
||||||
if session_id not in self.keys:
|
if device_type_upper == "ANDROID":
|
||||||
self.keys[session_id] = []
|
return DeviceTypes.ANDROID
|
||||||
session_keys = self.keys[session_id]
|
elif device_type_upper == "CHROME":
|
||||||
|
return DeviceTypes.CHROME
|
||||||
|
else:
|
||||||
|
return DeviceTypes.CHROME
|
||||||
|
|
||||||
for cached_key in res.get("cached_keys", []):
|
@property
|
||||||
# Handle KID format - could be hex string or UUID string
|
def is_playready(self) -> bool:
|
||||||
kid_str = cached_key["kid"]
|
"""Check if this CDM is in PlayReady mode."""
|
||||||
|
return self._is_playready
|
||||||
|
|
||||||
|
@property
|
||||||
|
def certificate_chain(self) -> MockCertificateChain:
|
||||||
|
"""Mock certificate chain for PlayReady compatibility."""
|
||||||
|
return MockCertificateChain(f"{self.device_name}_Remote")
|
||||||
|
|
||||||
|
def set_pssh_b64(self, pssh_b64: str) -> None:
|
||||||
|
"""Store base64-encoded PSSH data for PlayReady compatibility."""
|
||||||
|
self._pssh_b64 = pssh_b64
|
||||||
|
|
||||||
|
def set_required_kids(self, kids: List[Union[str, UUID]]) -> None:
|
||||||
|
"""
|
||||||
|
Set the required Key IDs for intelligent caching decisions.
|
||||||
|
|
||||||
|
This method enables the CDM to make smart decisions about when to request
|
||||||
|
additional keys via license challenges. When cached keys are available,
|
||||||
|
the CDM will compare them against the required KIDs to determine if a
|
||||||
|
license request is still needed for missing keys.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
kids: List of required Key IDs as UUIDs or hex strings
|
||||||
|
|
||||||
|
Note:
|
||||||
|
Should be called by DRM classes (PlayReady/Widevine) before making
|
||||||
|
license challenge requests to enable optimal caching behavior.
|
||||||
|
"""
|
||||||
|
self._required_kids = []
|
||||||
|
for kid in kids:
|
||||||
|
if isinstance(kid, UUID):
|
||||||
|
self._required_kids.append(str(kid).replace("-", "").lower())
|
||||||
|
else:
|
||||||
|
self._required_kids.append(str(kid).replace("-", "").lower())
|
||||||
|
|
||||||
|
def _generate_session_id(self) -> bytes:
|
||||||
|
"""Generate a unique session ID."""
|
||||||
|
return secrets.token_bytes(16)
|
||||||
|
|
||||||
|
def _get_init_data_from_pssh(self, pssh: Any) -> str:
|
||||||
|
"""Extract init data from various PSSH formats."""
|
||||||
|
if self.is_playready and self._pssh_b64:
|
||||||
|
return self._pssh_b64
|
||||||
|
|
||||||
|
if hasattr(pssh, "dumps"):
|
||||||
|
dumps_result = pssh.dumps()
|
||||||
|
|
||||||
|
if isinstance(dumps_result, str):
|
||||||
try:
|
try:
|
||||||
# Try as UUID string first
|
base64.b64decode(dumps_result)
|
||||||
kid_uuid = UUID(kid_str)
|
return dumps_result
|
||||||
except ValueError:
|
except Exception:
|
||||||
try:
|
return base64.b64encode(dumps_result.encode("utf-8")).decode("utf-8")
|
||||||
# Try as hex string (like the existing code)
|
else:
|
||||||
kid_uuid = UUID(bytes=bytes.fromhex(kid_str))
|
return base64.b64encode(dumps_result).decode("utf-8")
|
||||||
except ValueError:
|
elif hasattr(pssh, "raw"):
|
||||||
# Fallback: use Key.kid_to_uuid
|
raw_data = pssh.raw
|
||||||
kid_uuid = Key.kid_to_uuid(kid_str)
|
if isinstance(raw_data, str):
|
||||||
|
raw_data = raw_data.encode("utf-8")
|
||||||
|
return base64.b64encode(raw_data).decode("utf-8")
|
||||||
|
elif hasattr(pssh, "__class__") and "WrmHeader" in pssh.__class__.__name__:
|
||||||
|
if self.is_playready:
|
||||||
|
raise ValueError("PlayReady WRM header received but no PSSH B64 was set via set_pssh_b64()")
|
||||||
|
|
||||||
session_keys.append(Key(kid=kid_uuid, type_="CONTENT", key=bytes.fromhex(cached_key["key"])))
|
if hasattr(pssh, "raw_bytes"):
|
||||||
|
return base64.b64encode(pssh.raw_bytes).decode("utf-8")
|
||||||
|
elif hasattr(pssh, "bytes"):
|
||||||
|
return base64.b64encode(pssh.bytes).decode("utf-8")
|
||||||
|
else:
|
||||||
|
raise ValueError(f"Cannot extract PSSH data from WRM header type: {type(pssh)}")
|
||||||
|
else:
|
||||||
|
raise ValueError(f"Unsupported PSSH type: {type(pssh)}")
|
||||||
|
|
||||||
# Return empty challenge since we already have the keys
|
def open(self) -> bytes:
|
||||||
self.license_request = ""
|
"""
|
||||||
self.api_session_ids[session_id] = None
|
Open a new CDM session.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Session identifier as bytes
|
||||||
|
"""
|
||||||
|
session_id = self._generate_session_id()
|
||||||
|
self._sessions[session_id] = {
|
||||||
|
"service_certificate": None,
|
||||||
|
"keys": [],
|
||||||
|
"pssh": None,
|
||||||
|
"challenge": None,
|
||||||
|
"decrypt_labs_session_id": None,
|
||||||
|
"tried_cache": False,
|
||||||
|
"cached_keys": None,
|
||||||
|
}
|
||||||
|
return session_id
|
||||||
|
|
||||||
|
def close(self, session_id: bytes) -> None:
|
||||||
|
"""
|
||||||
|
Close a CDM session and perform comprehensive cleanup.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
session_id: Session identifier
|
||||||
|
|
||||||
|
Raises:
|
||||||
|
ValueError: If session ID is invalid
|
||||||
|
"""
|
||||||
|
if session_id not in self._sessions:
|
||||||
|
raise DecryptLabsRemoteCDMExceptions.InvalidSession(f"Invalid session ID: {session_id.hex()}")
|
||||||
|
|
||||||
|
session = self._sessions[session_id]
|
||||||
|
session.clear()
|
||||||
|
del self._sessions[session_id]
|
||||||
|
|
||||||
|
def get_service_certificate(self, session_id: bytes) -> Optional[bytes]:
|
||||||
|
"""
|
||||||
|
Get the service certificate for a session.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
session_id: Session identifier
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Service certificate if set, None otherwise
|
||||||
|
|
||||||
|
Raises:
|
||||||
|
ValueError: If session ID is invalid
|
||||||
|
"""
|
||||||
|
if session_id not in self._sessions:
|
||||||
|
raise DecryptLabsRemoteCDMExceptions.InvalidSession(f"Invalid session ID: {session_id.hex()}")
|
||||||
|
|
||||||
|
return self._sessions[session_id]["service_certificate"]
|
||||||
|
|
||||||
|
def set_service_certificate(self, session_id: bytes, certificate: Optional[Union[bytes, str]]) -> str:
|
||||||
|
"""
|
||||||
|
Set the service certificate for a session.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
session_id: Session identifier
|
||||||
|
certificate: Service certificate (bytes or base64 string)
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Certificate status message
|
||||||
|
|
||||||
|
Raises:
|
||||||
|
ValueError: If session ID is invalid
|
||||||
|
"""
|
||||||
|
if session_id not in self._sessions:
|
||||||
|
raise DecryptLabsRemoteCDMExceptions.InvalidSession(f"Invalid session ID: {session_id.hex()}")
|
||||||
|
|
||||||
|
if certificate is None:
|
||||||
|
if not self._is_playready and self.device_name == "L1":
|
||||||
|
certificate = WidevineCdm.common_privacy_cert
|
||||||
|
self._sessions[session_id]["service_certificate"] = base64.b64decode(certificate)
|
||||||
|
return "Using default Widevine common privacy certificate for L1"
|
||||||
|
else:
|
||||||
|
self._sessions[session_id]["service_certificate"] = None
|
||||||
|
return "No certificate set (not required for this device type)"
|
||||||
|
|
||||||
|
if isinstance(certificate, str):
|
||||||
|
certificate = base64.b64decode(certificate)
|
||||||
|
|
||||||
|
self._sessions[session_id]["service_certificate"] = certificate
|
||||||
|
return "Successfully set Service Certificate"
|
||||||
|
|
||||||
|
def has_cached_keys(self, session_id: bytes) -> bool:
|
||||||
|
"""
|
||||||
|
Check if cached keys are available for the session.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
session_id: Session identifier
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
True if cached keys are available
|
||||||
|
|
||||||
|
Raises:
|
||||||
|
ValueError: If session ID is invalid
|
||||||
|
"""
|
||||||
|
if session_id not in self._sessions:
|
||||||
|
raise DecryptLabsRemoteCDMExceptions.InvalidSession(f"Invalid session ID: {session_id.hex()}")
|
||||||
|
|
||||||
|
session = self._sessions[session_id]
|
||||||
|
session_keys = session.get("keys", [])
|
||||||
|
return len(session_keys) > 0
|
||||||
|
|
||||||
|
def get_license_challenge(
|
||||||
|
self, session_id: bytes, pssh_or_wrm: Any, license_type: str = "STREAMING", privacy_mode: bool = True
|
||||||
|
) -> bytes:
|
||||||
|
"""
|
||||||
|
Generate a license challenge using Decrypt Labs API with intelligent caching.
|
||||||
|
|
||||||
|
This method implements smart caching logic that:
|
||||||
|
1. First checks local vaults for required keys
|
||||||
|
2. Attempts to retrieve cached keys from the API
|
||||||
|
3. If required KIDs are set, compares available keys (vault + cached) against requirements
|
||||||
|
4. Only makes a license request if keys are missing
|
||||||
|
5. Returns empty challenge if all required keys are available
|
||||||
|
|
||||||
|
The intelligent caching works as follows:
|
||||||
|
- Local vaults: Always checked first if available
|
||||||
|
- For L1/L2 devices: Always prioritizes cached keys (API automatically optimizes)
|
||||||
|
- For other devices: Uses cache retry logic based on session state
|
||||||
|
- With required KIDs set: Only requests license for missing keys
|
||||||
|
- Without required KIDs: Returns any available cached keys
|
||||||
|
- For PlayReady: Combines vault, cached, and license keys seamlessly
|
||||||
|
|
||||||
|
Args:
|
||||||
|
session_id: Session identifier
|
||||||
|
pssh_or_wrm: PSSH object or WRM header (for PlayReady compatibility)
|
||||||
|
license_type: Type of license (STREAMING, OFFLINE, AUTOMATIC) - for compatibility only
|
||||||
|
privacy_mode: Whether to use privacy mode - for compatibility only
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
License challenge as bytes, or empty bytes if available keys satisfy requirements
|
||||||
|
|
||||||
|
Raises:
|
||||||
|
InvalidSession: If session ID is invalid
|
||||||
|
requests.RequestException: If API request fails
|
||||||
|
|
||||||
|
Note:
|
||||||
|
Call set_required_kids() before this method for optimal caching behavior.
|
||||||
|
L1/L2 devices automatically use cached keys when available per API design.
|
||||||
|
Local vault keys are always checked first when vaults are available.
|
||||||
|
"""
|
||||||
|
_ = license_type, privacy_mode
|
||||||
|
|
||||||
|
if session_id not in self._sessions:
|
||||||
|
raise DecryptLabsRemoteCDMExceptions.InvalidSession(f"Invalid session ID: {session_id.hex()}")
|
||||||
|
|
||||||
|
session = self._sessions[session_id]
|
||||||
|
|
||||||
|
session["pssh"] = pssh_or_wrm
|
||||||
|
init_data = self._get_init_data_from_pssh(pssh_or_wrm)
|
||||||
|
already_tried_cache = session.get("tried_cache", False)
|
||||||
|
|
||||||
|
if self.vaults and self._required_kids:
|
||||||
|
vault_keys = []
|
||||||
|
for kid_str in self._required_kids:
|
||||||
|
try:
|
||||||
|
clean_kid = kid_str.replace("-", "")
|
||||||
|
if len(clean_kid) == 32:
|
||||||
|
kid_uuid = UUID(hex=clean_kid)
|
||||||
|
else:
|
||||||
|
kid_uuid = UUID(hex=clean_kid.ljust(32, "0"))
|
||||||
|
key, _ = self.vaults.get_key(kid_uuid)
|
||||||
|
if key and key.count("0") != len(key):
|
||||||
|
vault_keys.append({"kid": kid_str, "key": key, "type": "CONTENT"})
|
||||||
|
except (ValueError, TypeError):
|
||||||
|
continue
|
||||||
|
|
||||||
|
if vault_keys:
|
||||||
|
vault_kids = set(k["kid"] for k in vault_keys)
|
||||||
|
required_kids = set(self._required_kids)
|
||||||
|
|
||||||
|
if required_kids.issubset(vault_kids):
|
||||||
|
session["keys"] = vault_keys
|
||||||
|
return b""
|
||||||
|
else:
|
||||||
|
session["vault_keys"] = vault_keys
|
||||||
|
|
||||||
|
if self.device_name in ["L1", "L2"]:
|
||||||
|
get_cached_keys = True
|
||||||
|
else:
|
||||||
|
get_cached_keys = not already_tried_cache
|
||||||
|
|
||||||
|
request_data = {
|
||||||
|
"scheme": self.device_name,
|
||||||
|
"init_data": init_data,
|
||||||
|
"get_cached_keys_if_exists": get_cached_keys,
|
||||||
|
}
|
||||||
|
|
||||||
|
if self.service_name:
|
||||||
|
request_data["service"] = self.service_name
|
||||||
|
|
||||||
|
if session["service_certificate"]:
|
||||||
|
request_data["service_certificate"] = base64.b64encode(session["service_certificate"]).decode("utf-8")
|
||||||
|
|
||||||
|
response = self._http_session.post(f"{self.host}/get-request", json=request_data, timeout=30)
|
||||||
|
|
||||||
|
if response.status_code != 200:
|
||||||
|
raise requests.RequestException(f"API request failed: {response.status_code} {response.text}")
|
||||||
|
|
||||||
|
data = response.json()
|
||||||
|
|
||||||
|
if data.get("message") != "success":
|
||||||
|
error_msg = data.get("message", "Unknown error")
|
||||||
|
if "details" in data:
|
||||||
|
error_msg += f" - Details: {data['details']}"
|
||||||
|
if "error" in data:
|
||||||
|
error_msg += f" - Error: {data['error']}"
|
||||||
|
|
||||||
|
if "service_certificate is required" in str(data) and not session["service_certificate"]:
|
||||||
|
error_msg += " (No service certificate was provided to the CDM session)"
|
||||||
|
|
||||||
|
raise requests.RequestException(f"API error: {error_msg}")
|
||||||
|
|
||||||
|
message_type = data.get("message_type")
|
||||||
|
|
||||||
|
if message_type == "cached-keys" or "cached_keys" in data:
|
||||||
|
"""
|
||||||
|
Handle cached keys response from API.
|
||||||
|
|
||||||
|
When the API returns cached keys, we need to determine if they satisfy
|
||||||
|
our requirements or if we need to make an additional license request
|
||||||
|
for missing keys.
|
||||||
|
"""
|
||||||
|
cached_keys = data.get("cached_keys", [])
|
||||||
|
parsed_keys = self._parse_cached_keys(cached_keys)
|
||||||
|
|
||||||
|
all_available_keys = list(parsed_keys)
|
||||||
|
if "vault_keys" in session:
|
||||||
|
all_available_keys.extend(session["vault_keys"])
|
||||||
|
|
||||||
|
session["keys"] = all_available_keys
|
||||||
|
session["tried_cache"] = True
|
||||||
|
|
||||||
|
if self._required_kids:
|
||||||
|
available_kids = set()
|
||||||
|
for key in all_available_keys:
|
||||||
|
if isinstance(key, dict) and "kid" in key:
|
||||||
|
available_kids.add(key["kid"].replace("-", "").lower())
|
||||||
|
|
||||||
|
required_kids = set(self._required_kids)
|
||||||
|
missing_kids = required_kids - available_kids
|
||||||
|
|
||||||
|
if missing_kids:
|
||||||
|
session["cached_keys"] = parsed_keys
|
||||||
|
|
||||||
|
if self.device_name in ["L1", "L2"]:
|
||||||
|
license_request_data = {
|
||||||
|
"scheme": self.device_name,
|
||||||
|
"init_data": init_data,
|
||||||
|
"get_cached_keys_if_exists": False,
|
||||||
|
}
|
||||||
|
if self.service_name:
|
||||||
|
license_request_data["service"] = self.service_name
|
||||||
|
if session["service_certificate"]:
|
||||||
|
license_request_data["service_certificate"] = base64.b64encode(
|
||||||
|
session["service_certificate"]
|
||||||
|
).decode("utf-8")
|
||||||
|
else:
|
||||||
|
license_request_data = request_data.copy()
|
||||||
|
license_request_data["get_cached_keys_if_exists"] = False
|
||||||
|
|
||||||
|
session["decrypt_labs_session_id"] = None
|
||||||
|
session["challenge"] = None
|
||||||
|
session["tried_cache"] = False
|
||||||
|
|
||||||
|
response = self._http_session.post(
|
||||||
|
f"{self.host}/get-request", json=license_request_data, timeout=30
|
||||||
|
)
|
||||||
|
if response.status_code == 200:
|
||||||
|
data = response.json()
|
||||||
|
if data.get("message") == "success" and "challenge" in data:
|
||||||
|
challenge = base64.b64decode(data["challenge"])
|
||||||
|
session["challenge"] = challenge
|
||||||
|
session["decrypt_labs_session_id"] = data["session_id"]
|
||||||
|
return challenge
|
||||||
|
|
||||||
|
return b""
|
||||||
|
else:
|
||||||
|
return b""
|
||||||
|
else:
|
||||||
|
return b""
|
||||||
|
|
||||||
|
if message_type == "license-request" or "challenge" in data:
|
||||||
|
challenge = base64.b64decode(data["challenge"])
|
||||||
|
session["challenge"] = challenge
|
||||||
|
session["decrypt_labs_session_id"] = data["session_id"]
|
||||||
|
return challenge
|
||||||
|
|
||||||
|
error_msg = f"Unexpected API response format. message_type={message_type}, available_fields={list(data.keys())}"
|
||||||
|
if data.get("message"):
|
||||||
|
error_msg = f"API response: {data['message']} - {error_msg}"
|
||||||
|
if "details" in data:
|
||||||
|
error_msg += f" - Details: {data['details']}"
|
||||||
|
if "error" in data:
|
||||||
|
error_msg += f" - Error: {data['error']}"
|
||||||
|
|
||||||
|
if already_tried_cache and data.get("message") == "success":
|
||||||
return b""
|
return b""
|
||||||
|
|
||||||
# Normal challenge response
|
raise requests.RequestException(error_msg)
|
||||||
self.license_request = res["challenge"]
|
|
||||||
self.api_session_ids[session_id] = res["session_id"]
|
|
||||||
|
|
||||||
return base64.b64decode(self.license_request)
|
def parse_license(self, session_id: bytes, license_message: Union[bytes, str]) -> None:
|
||||||
|
"""
|
||||||
|
Parse license response using Decrypt Labs API with intelligent key combination.
|
||||||
|
|
||||||
def parse_license(self, session_id: bytes, license_message: Union[SignedMessage, bytes, str]) -> None:
|
For PlayReady content with partial cached keys, this method intelligently
|
||||||
session_id_api = self.api_session_ids[session_id]
|
combines the cached keys with newly obtained license keys, avoiding
|
||||||
if session_id not in self.keys:
|
duplicates while ensuring all required keys are available.
|
||||||
self.keys[session_id] = []
|
|
||||||
session_keys = self.keys[session_id]
|
|
||||||
|
|
||||||
# If we already have cached keys and no session_id_api, skip processing
|
The key combination process:
|
||||||
if session_id_api is None and session_keys:
|
1. Extracts keys from the license response
|
||||||
|
2. If cached keys exist (PlayReady), combines them with license keys
|
||||||
|
3. Removes duplicate keys by comparing normalized KIDs
|
||||||
|
4. Updates the session with the complete key set
|
||||||
|
|
||||||
|
Args:
|
||||||
|
session_id: Session identifier
|
||||||
|
license_message: License response from license server
|
||||||
|
|
||||||
|
Raises:
|
||||||
|
ValueError: If session ID is invalid or no challenge available
|
||||||
|
requests.RequestException: If API request fails
|
||||||
|
"""
|
||||||
|
if session_id not in self._sessions:
|
||||||
|
raise DecryptLabsRemoteCDMExceptions.InvalidSession(f"Invalid session ID: {session_id.hex()}")
|
||||||
|
|
||||||
|
session = self._sessions[session_id]
|
||||||
|
|
||||||
|
if session["keys"] and not (self.is_playready and "cached_keys" in session):
|
||||||
return
|
return
|
||||||
|
|
||||||
if isinstance(license_message, dict) and "keys" in license_message:
|
if not session.get("challenge") or not session.get("decrypt_labs_session_id"):
|
||||||
session_keys.extend(
|
raise ValueError("No challenge available - call get_license_challenge first")
|
||||||
[
|
|
||||||
Key(kid=Key.kid_to_uuid(x["kid"]), type_=x.get("type", "CONTENT"), key=bytes.fromhex(x["key"]))
|
|
||||||
for x in license_message["keys"]
|
|
||||||
]
|
|
||||||
)
|
|
||||||
|
|
||||||
else:
|
if isinstance(license_message, str):
|
||||||
# Ensure license_message is base64 encoded
|
if self.is_playready and license_message.strip().startswith("<?xml"):
|
||||||
if isinstance(license_message, bytes):
|
license_message = license_message.encode("utf-8")
|
||||||
license_response_b64 = base64.b64encode(license_message).decode()
|
|
||||||
elif isinstance(license_message, str):
|
|
||||||
license_response_b64 = license_message
|
|
||||||
else:
|
else:
|
||||||
license_response_b64 = str(license_message)
|
try:
|
||||||
res = self.session(
|
license_message = base64.b64decode(license_message)
|
||||||
self.host + "/decrypt-response",
|
except Exception:
|
||||||
{
|
license_message = license_message.encode("utf-8")
|
||||||
"session_id": session_id_api,
|
|
||||||
"init_data": self.pssh.dumps(),
|
|
||||||
"license_request": self.license_request,
|
|
||||||
"license_response": license_response_b64,
|
|
||||||
"scheme": self.scheme,
|
|
||||||
},
|
|
||||||
)
|
|
||||||
|
|
||||||
original_keys = res["keys"].replace("\n", " ")
|
pssh = session["pssh"]
|
||||||
keys_separated = original_keys.split("--key ")
|
init_data = self._get_init_data_from_pssh(pssh)
|
||||||
formatted_keys = []
|
|
||||||
for k in keys_separated:
|
license_request_b64 = base64.b64encode(session["challenge"]).decode("utf-8")
|
||||||
if ":" in k:
|
license_response_b64 = base64.b64encode(license_message).decode("utf-8")
|
||||||
key = k.strip()
|
|
||||||
formatted_keys.append(key)
|
request_data = {
|
||||||
for keys in formatted_keys:
|
"scheme": self.device_name,
|
||||||
session_keys.append(
|
"session_id": session["decrypt_labs_session_id"],
|
||||||
(
|
"init_data": init_data,
|
||||||
Key(
|
"license_request": license_request_b64,
|
||||||
kid=UUID(bytes=bytes.fromhex(keys.split(":")[0])),
|
"license_response": license_response_b64,
|
||||||
type_="CONTENT",
|
}
|
||||||
key=bytes.fromhex(keys.split(":")[1]),
|
|
||||||
)
|
response = self._http_session.post(f"{self.host}/decrypt-response", json=request_data, timeout=30)
|
||||||
)
|
|
||||||
|
if response.status_code != 200:
|
||||||
|
raise requests.RequestException(f"License decrypt failed: {response.status_code} {response.text}")
|
||||||
|
|
||||||
|
data = response.json()
|
||||||
|
|
||||||
|
if data.get("message") != "success":
|
||||||
|
error_msg = data.get("message", "Unknown error")
|
||||||
|
if "error" in data:
|
||||||
|
error_msg += f" - Error: {data['error']}"
|
||||||
|
if "details" in data:
|
||||||
|
error_msg += f" - Details: {data['details']}"
|
||||||
|
raise requests.RequestException(f"License decrypt error: {error_msg}")
|
||||||
|
|
||||||
|
license_keys = self._parse_keys_response(data)
|
||||||
|
|
||||||
|
all_keys = []
|
||||||
|
|
||||||
|
if "vault_keys" in session:
|
||||||
|
all_keys.extend(session["vault_keys"])
|
||||||
|
|
||||||
|
if "cached_keys" in session:
|
||||||
|
cached_keys = session.get("cached_keys", [])
|
||||||
|
if cached_keys:
|
||||||
|
for cached_key in cached_keys:
|
||||||
|
all_keys.append(cached_key)
|
||||||
|
|
||||||
|
for license_key in license_keys:
|
||||||
|
already_exists = False
|
||||||
|
license_kid = None
|
||||||
|
if isinstance(license_key, dict) and "kid" in license_key:
|
||||||
|
license_kid = license_key["kid"].replace("-", "").lower()
|
||||||
|
elif hasattr(license_key, "kid"):
|
||||||
|
license_kid = str(license_key.kid).replace("-", "").lower()
|
||||||
|
elif hasattr(license_key, "key_id"):
|
||||||
|
license_kid = str(license_key.key_id).replace("-", "").lower()
|
||||||
|
|
||||||
|
if license_kid:
|
||||||
|
for existing_key in all_keys:
|
||||||
|
existing_kid = None
|
||||||
|
if isinstance(existing_key, dict) and "kid" in existing_key:
|
||||||
|
existing_kid = existing_key["kid"].replace("-", "").lower()
|
||||||
|
elif hasattr(existing_key, "kid"):
|
||||||
|
existing_kid = str(existing_key.kid).replace("-", "").lower()
|
||||||
|
elif hasattr(existing_key, "key_id"):
|
||||||
|
existing_kid = str(existing_key.key_id).replace("-", "").lower()
|
||||||
|
|
||||||
|
if existing_kid == license_kid:
|
||||||
|
already_exists = True
|
||||||
|
break
|
||||||
|
|
||||||
|
if not already_exists:
|
||||||
|
all_keys.append(license_key)
|
||||||
|
|
||||||
|
session["keys"] = all_keys
|
||||||
|
session.pop("cached_keys", None)
|
||||||
|
session.pop("vault_keys", None)
|
||||||
|
|
||||||
|
if self.vaults and session["keys"]:
|
||||||
|
key_dict = {}
|
||||||
|
for key in session["keys"]:
|
||||||
|
if key["type"] == "CONTENT":
|
||||||
|
try:
|
||||||
|
clean_kid = key["kid"].replace("-", "")
|
||||||
|
if len(clean_kid) == 32:
|
||||||
|
kid_uuid = UUID(hex=clean_kid)
|
||||||
|
else:
|
||||||
|
kid_uuid = UUID(hex=clean_kid.ljust(32, "0"))
|
||||||
|
key_dict[kid_uuid] = key["key"]
|
||||||
|
except (ValueError, TypeError):
|
||||||
|
continue
|
||||||
|
if key_dict:
|
||||||
|
self.vaults.add_keys(key_dict)
|
||||||
|
|
||||||
|
def get_keys(self, session_id: bytes, type_: Optional[str] = None) -> List[Key]:
|
||||||
|
"""
|
||||||
|
Get keys from the session.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
session_id: Session identifier
|
||||||
|
type_: Optional key type filter (CONTENT, SIGNING, etc.)
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
List of Key objects
|
||||||
|
|
||||||
|
Raises:
|
||||||
|
InvalidSession: If session ID is invalid
|
||||||
|
"""
|
||||||
|
if session_id not in self._sessions:
|
||||||
|
raise DecryptLabsRemoteCDMExceptions.InvalidSession(f"Invalid session ID: {session_id.hex()}")
|
||||||
|
|
||||||
|
key_dicts = self._sessions[session_id]["keys"]
|
||||||
|
keys = [Key(kid=k["kid"], key=k["key"], type_=k["type"]) for k in key_dicts]
|
||||||
|
|
||||||
|
if type_:
|
||||||
|
keys = [key for key in keys if key.type == type_]
|
||||||
|
|
||||||
|
return keys
|
||||||
|
|
||||||
|
def _parse_cached_keys(self, cached_keys_data: List[Dict[str, Any]]) -> List[Dict[str, Any]]:
|
||||||
|
"""Parse cached keys from API response.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
cached_keys_data: List of cached key objects from API
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
List of key dictionaries
|
||||||
|
"""
|
||||||
|
keys = []
|
||||||
|
|
||||||
|
try:
|
||||||
|
if cached_keys_data and isinstance(cached_keys_data, list):
|
||||||
|
for key_data in cached_keys_data:
|
||||||
|
if "kid" in key_data and "key" in key_data:
|
||||||
|
keys.append({"kid": key_data["kid"], "key": key_data["key"], "type": "CONTENT"})
|
||||||
|
except Exception:
|
||||||
|
pass
|
||||||
|
return keys
|
||||||
|
|
||||||
|
def _parse_keys_response(self, data: Dict[str, Any]) -> List[Dict[str, Any]]:
|
||||||
|
"""Parse keys from decrypt response."""
|
||||||
|
keys = []
|
||||||
|
|
||||||
|
if "keys" in data and isinstance(data["keys"], str):
|
||||||
|
keys_string = data["keys"]
|
||||||
|
|
||||||
|
for line in keys_string.split("\n"):
|
||||||
|
line = line.strip()
|
||||||
|
if line.startswith("--key "):
|
||||||
|
key_part = line[6:]
|
||||||
|
if ":" in key_part:
|
||||||
|
kid, key = key_part.split(":", 1)
|
||||||
|
keys.append({"kid": kid.strip(), "key": key.strip(), "type": "CONTENT"})
|
||||||
|
elif "keys" in data and isinstance(data["keys"], list):
|
||||||
|
for key_data in data["keys"]:
|
||||||
|
keys.append(
|
||||||
|
{"kid": key_data.get("kid"), "key": key_data.get("key"), "type": key_data.get("type", "CONTENT")}
|
||||||
)
|
)
|
||||||
|
|
||||||
def get_keys(self, session_id: bytes, type_: Optional[Union[int, str]] = None) -> list[Key]:
|
return keys
|
||||||
return self.keys[session_id]
|
|
||||||
|
|
||||||
def session(self, url, data, retries=3):
|
|
||||||
res = self.req_session.post(url, json=data).json()
|
|
||||||
|
|
||||||
if res.get("message") != "success":
|
__all__ = ["DecryptLabsRemoteCDM"]
|
||||||
if "License Response Decryption Process Failed at the very beginning" in res.get("Error", ""):
|
|
||||||
if retries > 0:
|
|
||||||
return self.session(url, data, retries=retries - 1)
|
|
||||||
else:
|
|
||||||
raise ValueError(f"CDM API returned an error: {res['Error']}")
|
|
||||||
else:
|
|
||||||
raise ValueError(f"CDM API returned an error: {res['Error']}")
|
|
||||||
|
|
||||||
return res
|
|
||||||
|
|||||||
@@ -88,6 +88,7 @@ class Config:
|
|||||||
self.tag_group_name: bool = kwargs.get("tag_group_name", True)
|
self.tag_group_name: bool = kwargs.get("tag_group_name", True)
|
||||||
self.tag_imdb_tmdb: bool = kwargs.get("tag_imdb_tmdb", True)
|
self.tag_imdb_tmdb: bool = kwargs.get("tag_imdb_tmdb", True)
|
||||||
self.tmdb_api_key: str = kwargs.get("tmdb_api_key") or ""
|
self.tmdb_api_key: str = kwargs.get("tmdb_api_key") or ""
|
||||||
|
self.decrypt_labs_api_key: str = kwargs.get("decrypt_labs_api_key") or ""
|
||||||
self.update_checks: bool = kwargs.get("update_checks", True)
|
self.update_checks: bool = kwargs.get("update_checks", True)
|
||||||
self.update_check_interval: int = kwargs.get("update_check_interval", 24)
|
self.update_check_interval: int = kwargs.get("update_check_interval", 24)
|
||||||
self.scene_naming: bool = kwargs.get("scene_naming", True)
|
self.scene_naming: bool = kwargs.get("scene_naming", True)
|
||||||
|
|||||||
@@ -150,6 +150,7 @@ def download(
|
|||||||
|
|
||||||
track_type = track.__class__.__name__
|
track_type = track.__class__.__name__
|
||||||
thread_count = str(config.n_m3u8dl_re.get("thread_count", max_workers))
|
thread_count = str(config.n_m3u8dl_re.get("thread_count", max_workers))
|
||||||
|
retry_count = str(config.n_m3u8dl_re.get("retry_count", max_workers))
|
||||||
ad_keyword = config.n_m3u8dl_re.get("ad_keyword")
|
ad_keyword = config.n_m3u8dl_re.get("ad_keyword")
|
||||||
|
|
||||||
arguments = [
|
arguments = [
|
||||||
@@ -160,6 +161,8 @@ def download(
|
|||||||
output_dir,
|
output_dir,
|
||||||
"--thread-count",
|
"--thread-count",
|
||||||
thread_count,
|
thread_count,
|
||||||
|
"--download-retry-count",
|
||||||
|
retry_count,
|
||||||
"--no-log",
|
"--no-log",
|
||||||
"--write-meta-json",
|
"--write-meta-json",
|
||||||
"false",
|
"false",
|
||||||
|
|||||||
@@ -8,6 +8,7 @@ from urllib.parse import urljoin
|
|||||||
|
|
||||||
from Cryptodome.Cipher import AES
|
from Cryptodome.Cipher import AES
|
||||||
from Cryptodome.Util.Padding import unpad
|
from Cryptodome.Util.Padding import unpad
|
||||||
|
from curl_cffi.requests import Session as CurlSession
|
||||||
from m3u8.model import Key
|
from m3u8.model import Key
|
||||||
from requests import Session
|
from requests import Session
|
||||||
|
|
||||||
@@ -69,8 +70,8 @@ class ClearKey:
|
|||||||
"""
|
"""
|
||||||
if not isinstance(m3u_key, Key):
|
if not isinstance(m3u_key, Key):
|
||||||
raise ValueError(f"Provided M3U Key is in an unexpected type {m3u_key!r}")
|
raise ValueError(f"Provided M3U Key is in an unexpected type {m3u_key!r}")
|
||||||
if not isinstance(session, (Session, type(None))):
|
if not isinstance(session, (Session, CurlSession, type(None))):
|
||||||
raise TypeError(f"Expected session to be a {Session}, not a {type(session)}")
|
raise TypeError(f"Expected session to be a {Session} or {CurlSession}, not a {type(session)}")
|
||||||
|
|
||||||
if not m3u_key.method.startswith("AES"):
|
if not m3u_key.method.startswith("AES"):
|
||||||
raise ValueError(f"Provided M3U Key is not an AES Clear Key, {m3u_key.method}")
|
raise ValueError(f"Provided M3U Key is not an AES Clear Key, {m3u_key.method}")
|
||||||
|
|||||||
@@ -224,31 +224,70 @@ class PlayReady:
|
|||||||
def kids(self) -> list[UUID]:
|
def kids(self) -> list[UUID]:
|
||||||
return self._kids
|
return self._kids
|
||||||
|
|
||||||
def get_content_keys(self, cdm: PlayReadyCdm, certificate: Callable, licence: Callable) -> None:
|
def _extract_keys_from_cdm(self, cdm: PlayReadyCdm, session_id: bytes) -> dict:
|
||||||
for kid in self.kids:
|
"""Extract keys from CDM session with cross-library compatibility.
|
||||||
if kid in self.content_keys:
|
|
||||||
|
Args:
|
||||||
|
cdm: CDM instance
|
||||||
|
session_id: Session identifier
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Dictionary mapping KID UUIDs to hex keys
|
||||||
|
"""
|
||||||
|
keys = {}
|
||||||
|
for key in cdm.get_keys(session_id):
|
||||||
|
if hasattr(key, "key_id"):
|
||||||
|
kid = key.key_id
|
||||||
|
elif hasattr(key, "kid"):
|
||||||
|
kid = key.kid
|
||||||
|
else:
|
||||||
continue
|
continue
|
||||||
session_id = cdm.open()
|
|
||||||
try:
|
|
||||||
challenge = cdm.get_license_challenge(session_id, self.pssh.wrm_headers[0])
|
|
||||||
license_res = licence(challenge=challenge)
|
|
||||||
|
|
||||||
if isinstance(license_res, bytes):
|
if hasattr(key, "key") and hasattr(key.key, "hex"):
|
||||||
license_str = license_res.decode(errors="ignore")
|
key_hex = key.key.hex()
|
||||||
else:
|
elif hasattr(key, "key") and isinstance(key.key, bytes):
|
||||||
license_str = str(license_res)
|
key_hex = key.key.hex()
|
||||||
|
elif hasattr(key, "key") and isinstance(key.key, str):
|
||||||
|
key_hex = key.key
|
||||||
|
else:
|
||||||
|
continue
|
||||||
|
|
||||||
if "<License>" not in license_str:
|
keys[kid] = key_hex
|
||||||
try:
|
return keys
|
||||||
license_str = base64.b64decode(license_str + "===").decode()
|
|
||||||
except Exception:
|
|
||||||
pass
|
|
||||||
|
|
||||||
cdm.parse_license(session_id, license_str)
|
def get_content_keys(self, cdm: PlayReadyCdm, certificate: Callable, licence: Callable) -> None:
|
||||||
keys = {key.key_id: key.key.hex() for key in cdm.get_keys(session_id)}
|
session_id = cdm.open()
|
||||||
self.content_keys.update(keys)
|
try:
|
||||||
finally:
|
if hasattr(cdm, "set_pssh_b64") and self.pssh_b64:
|
||||||
cdm.close(session_id)
|
cdm.set_pssh_b64(self.pssh_b64)
|
||||||
|
|
||||||
|
if hasattr(cdm, "set_required_kids"):
|
||||||
|
cdm.set_required_kids(self.kids)
|
||||||
|
|
||||||
|
challenge = cdm.get_license_challenge(session_id, self.pssh.wrm_headers[0])
|
||||||
|
|
||||||
|
if challenge:
|
||||||
|
try:
|
||||||
|
license_res = licence(challenge=challenge)
|
||||||
|
if isinstance(license_res, bytes):
|
||||||
|
license_str = license_res.decode(errors="ignore")
|
||||||
|
else:
|
||||||
|
license_str = str(license_res)
|
||||||
|
|
||||||
|
if "<License>" not in license_str:
|
||||||
|
try:
|
||||||
|
license_str = base64.b64decode(license_str + "===").decode()
|
||||||
|
except Exception:
|
||||||
|
pass
|
||||||
|
|
||||||
|
cdm.parse_license(session_id, license_str)
|
||||||
|
except Exception:
|
||||||
|
raise
|
||||||
|
|
||||||
|
keys = self._extract_keys_from_cdm(cdm, session_id)
|
||||||
|
self.content_keys.update(keys)
|
||||||
|
finally:
|
||||||
|
cdm.close(session_id)
|
||||||
|
|
||||||
if not self.content_keys:
|
if not self.content_keys:
|
||||||
raise PlayReady.Exceptions.EmptyLicense("No Content Keys were within the License")
|
raise PlayReady.Exceptions.EmptyLicense("No Content Keys were within the License")
|
||||||
|
|||||||
@@ -185,7 +185,15 @@ class Widevine:
|
|||||||
if cert and hasattr(cdm, "set_service_certificate"):
|
if cert and hasattr(cdm, "set_service_certificate"):
|
||||||
cdm.set_service_certificate(session_id, cert)
|
cdm.set_service_certificate(session_id, cert)
|
||||||
|
|
||||||
cdm.parse_license(session_id, licence(challenge=cdm.get_license_challenge(session_id, self.pssh)))
|
if hasattr(cdm, "set_required_kids"):
|
||||||
|
cdm.set_required_kids(self.kids)
|
||||||
|
|
||||||
|
challenge = cdm.get_license_challenge(session_id, self.pssh)
|
||||||
|
|
||||||
|
if hasattr(cdm, "has_cached_keys") and cdm.has_cached_keys(session_id):
|
||||||
|
pass
|
||||||
|
else:
|
||||||
|
cdm.parse_license(session_id, licence(challenge=challenge))
|
||||||
|
|
||||||
self.content_keys = {key.kid: key.key.hex() for key in cdm.get_keys(session_id, "CONTENT")}
|
self.content_keys = {key.kid: key.key.hex() for key in cdm.get_keys(session_id, "CONTENT")}
|
||||||
if not self.content_keys:
|
if not self.content_keys:
|
||||||
@@ -213,10 +221,18 @@ class Widevine:
|
|||||||
if cert and hasattr(cdm, "set_service_certificate"):
|
if cert and hasattr(cdm, "set_service_certificate"):
|
||||||
cdm.set_service_certificate(session_id, cert)
|
cdm.set_service_certificate(session_id, cert)
|
||||||
|
|
||||||
cdm.parse_license(
|
if hasattr(cdm, "set_required_kids"):
|
||||||
session_id,
|
cdm.set_required_kids(self.kids)
|
||||||
licence(session_id=session_id, challenge=cdm.get_license_challenge(session_id, self.pssh)),
|
|
||||||
)
|
challenge = cdm.get_license_challenge(session_id, self.pssh)
|
||||||
|
|
||||||
|
if hasattr(cdm, "has_cached_keys") and cdm.has_cached_keys(session_id):
|
||||||
|
pass
|
||||||
|
else:
|
||||||
|
cdm.parse_license(
|
||||||
|
session_id,
|
||||||
|
licence(session_id=session_id, challenge=challenge),
|
||||||
|
)
|
||||||
|
|
||||||
self.content_keys = {key.kid: key.key.hex() for key in cdm.get_keys(session_id, "CONTENT")}
|
self.content_keys = {key.kid: key.key.hex() for key in cdm.get_keys(session_id, "CONTENT")}
|
||||||
if not self.content_keys:
|
if not self.content_keys:
|
||||||
|
|||||||
@@ -15,6 +15,7 @@ from uuid import UUID
|
|||||||
from zlib import crc32
|
from zlib import crc32
|
||||||
|
|
||||||
import requests
|
import requests
|
||||||
|
from curl_cffi.requests import Session as CurlSession
|
||||||
from langcodes import Language, tag_is_valid
|
from langcodes import Language, tag_is_valid
|
||||||
from lxml.etree import Element, ElementTree
|
from lxml.etree import Element, ElementTree
|
||||||
from pyplayready.system.pssh import PSSH as PR_PSSH
|
from pyplayready.system.pssh import PSSH as PR_PSSH
|
||||||
@@ -47,7 +48,7 @@ class DASH:
|
|||||||
self.url = url
|
self.url = url
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def from_url(cls, url: str, session: Optional[Session] = None, **args: Any) -> DASH:
|
def from_url(cls, url: str, session: Optional[Union[Session, CurlSession]] = None, **args: Any) -> DASH:
|
||||||
if not url:
|
if not url:
|
||||||
raise requests.URLRequired("DASH manifest URL must be provided for relative path computations.")
|
raise requests.URLRequired("DASH manifest URL must be provided for relative path computations.")
|
||||||
if not isinstance(url, str):
|
if not isinstance(url, str):
|
||||||
@@ -55,8 +56,8 @@ class DASH:
|
|||||||
|
|
||||||
if not session:
|
if not session:
|
||||||
session = Session()
|
session = Session()
|
||||||
elif not isinstance(session, Session):
|
elif not isinstance(session, (Session, CurlSession)):
|
||||||
raise TypeError(f"Expected session to be a {Session}, not {session!r}")
|
raise TypeError(f"Expected session to be a {Session} or {CurlSession}, not {session!r}")
|
||||||
|
|
||||||
res = session.get(url, **args)
|
res = session.get(url, **args)
|
||||||
if res.url != url:
|
if res.url != url:
|
||||||
@@ -103,6 +104,10 @@ class DASH:
|
|||||||
continue
|
continue
|
||||||
if next(iter(period.xpath("SegmentType/@value")), "content") != "content":
|
if next(iter(period.xpath("SegmentType/@value")), "content") != "content":
|
||||||
continue
|
continue
|
||||||
|
if "urn:amazon:primevideo:cachingBreadth" in [
|
||||||
|
x.get("schemeIdUri") for x in period.findall("SupplementalProperty")
|
||||||
|
]:
|
||||||
|
continue
|
||||||
|
|
||||||
for adaptation_set in period.findall("AdaptationSet"):
|
for adaptation_set in period.findall("AdaptationSet"):
|
||||||
if self.is_trick_mode(adaptation_set):
|
if self.is_trick_mode(adaptation_set):
|
||||||
|
|||||||
@@ -14,9 +14,10 @@ from typing import Any, Callable, Optional, Union
|
|||||||
from urllib.parse import urljoin
|
from urllib.parse import urljoin
|
||||||
from zlib import crc32
|
from zlib import crc32
|
||||||
|
|
||||||
import httpx
|
|
||||||
import m3u8
|
import m3u8
|
||||||
import requests
|
import requests
|
||||||
|
from curl_cffi.requests import Response as CurlResponse
|
||||||
|
from curl_cffi.requests import Session as CurlSession
|
||||||
from langcodes import Language, tag_is_valid
|
from langcodes import Language, tag_is_valid
|
||||||
from m3u8 import M3U8
|
from m3u8 import M3U8
|
||||||
from pyplayready.cdm import Cdm as PlayReadyCdm
|
from pyplayready.cdm import Cdm as PlayReadyCdm
|
||||||
@@ -35,7 +36,7 @@ from unshackle.core.utilities import get_extension, is_close_match, try_ensure_u
|
|||||||
|
|
||||||
|
|
||||||
class HLS:
|
class HLS:
|
||||||
def __init__(self, manifest: M3U8, session: Optional[Union[Session, httpx.Client]] = None):
|
def __init__(self, manifest: M3U8, session: Optional[Union[Session, CurlSession]] = None):
|
||||||
if not manifest:
|
if not manifest:
|
||||||
raise ValueError("HLS manifest must be provided.")
|
raise ValueError("HLS manifest must be provided.")
|
||||||
if not isinstance(manifest, M3U8):
|
if not isinstance(manifest, M3U8):
|
||||||
@@ -47,7 +48,7 @@ class HLS:
|
|||||||
self.session = session or Session()
|
self.session = session or Session()
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def from_url(cls, url: str, session: Optional[Union[Session, httpx.Client]] = None, **args: Any) -> HLS:
|
def from_url(cls, url: str, session: Optional[Union[Session, CurlSession]] = None, **args: Any) -> HLS:
|
||||||
if not url:
|
if not url:
|
||||||
raise requests.URLRequired("HLS manifest URL must be provided.")
|
raise requests.URLRequired("HLS manifest URL must be provided.")
|
||||||
if not isinstance(url, str):
|
if not isinstance(url, str):
|
||||||
@@ -55,22 +56,22 @@ class HLS:
|
|||||||
|
|
||||||
if not session:
|
if not session:
|
||||||
session = Session()
|
session = Session()
|
||||||
elif not isinstance(session, (Session, httpx.Client)):
|
elif not isinstance(session, (Session, CurlSession)):
|
||||||
raise TypeError(f"Expected session to be a {Session} or {httpx.Client}, not {session!r}")
|
raise TypeError(f"Expected session to be a {Session} or {CurlSession}, not {session!r}")
|
||||||
|
|
||||||
res = session.get(url, **args)
|
res = session.get(url, **args)
|
||||||
|
|
||||||
# Handle both requests and httpx response objects
|
# Handle requests and curl_cffi response objects
|
||||||
if isinstance(res, requests.Response):
|
if isinstance(res, requests.Response):
|
||||||
if not res.ok:
|
if not res.ok:
|
||||||
raise requests.ConnectionError("Failed to request the M3U(8) document.", response=res)
|
raise requests.ConnectionError("Failed to request the M3U(8) document.", response=res)
|
||||||
content = res.text
|
content = res.text
|
||||||
elif isinstance(res, httpx.Response):
|
elif isinstance(res, CurlResponse):
|
||||||
if res.status_code >= 400:
|
if not res.ok:
|
||||||
raise requests.ConnectionError("Failed to request the M3U(8) document.", response=res)
|
raise requests.ConnectionError("Failed to request the M3U(8) document.", response=res)
|
||||||
content = res.text
|
content = res.text
|
||||||
else:
|
else:
|
||||||
raise TypeError(f"Expected response to be a requests.Response or httpx.Response, not {type(res)}")
|
raise TypeError(f"Expected response to be a requests.Response or curl_cffi.Response, not {type(res)}")
|
||||||
|
|
||||||
master = m3u8.loads(content, uri=url)
|
master = m3u8.loads(content, uri=url)
|
||||||
|
|
||||||
@@ -229,7 +230,7 @@ class HLS:
|
|||||||
save_path: Path,
|
save_path: Path,
|
||||||
save_dir: Path,
|
save_dir: Path,
|
||||||
progress: partial,
|
progress: partial,
|
||||||
session: Optional[Union[Session, httpx.Client]] = None,
|
session: Optional[Union[Session, CurlSession]] = None,
|
||||||
proxy: Optional[str] = None,
|
proxy: Optional[str] = None,
|
||||||
max_workers: Optional[int] = None,
|
max_workers: Optional[int] = None,
|
||||||
license_widevine: Optional[Callable] = None,
|
license_widevine: Optional[Callable] = None,
|
||||||
@@ -238,15 +239,13 @@ class HLS:
|
|||||||
) -> None:
|
) -> None:
|
||||||
if not session:
|
if not session:
|
||||||
session = Session()
|
session = Session()
|
||||||
elif not isinstance(session, (Session, httpx.Client)):
|
elif not isinstance(session, (Session, CurlSession)):
|
||||||
raise TypeError(f"Expected session to be a {Session} or {httpx.Client}, not {session!r}")
|
raise TypeError(f"Expected session to be a {Session} or {CurlSession}, not {session!r}")
|
||||||
|
|
||||||
if proxy:
|
if proxy:
|
||||||
# Handle proxies differently based on session type
|
# Handle proxies differently based on session type
|
||||||
if isinstance(session, Session):
|
if isinstance(session, Session):
|
||||||
session.proxies.update({"all": proxy})
|
session.proxies.update({"all": proxy})
|
||||||
elif isinstance(session, httpx.Client):
|
|
||||||
session.proxies = {"http://": proxy, "https://": proxy}
|
|
||||||
|
|
||||||
log = logging.getLogger("HLS")
|
log = logging.getLogger("HLS")
|
||||||
|
|
||||||
@@ -257,13 +256,8 @@ class HLS:
|
|||||||
log.error(f"Failed to request the invariant M3U8 playlist: {response.status_code}")
|
log.error(f"Failed to request the invariant M3U8 playlist: {response.status_code}")
|
||||||
sys.exit(1)
|
sys.exit(1)
|
||||||
playlist_text = response.text
|
playlist_text = response.text
|
||||||
elif isinstance(response, httpx.Response):
|
|
||||||
if response.status_code >= 400:
|
|
||||||
log.error(f"Failed to request the invariant M3U8 playlist: {response.status_code}")
|
|
||||||
sys.exit(1)
|
|
||||||
playlist_text = response.text
|
|
||||||
else:
|
else:
|
||||||
raise TypeError(f"Expected response to be a requests.Response or httpx.Response, not {type(response)}")
|
raise TypeError(f"Expected response to be a requests.Response or curl_cffi.Response, not {type(response)}")
|
||||||
|
|
||||||
master = m3u8.loads(playlist_text, uri=track.url)
|
master = m3u8.loads(playlist_text, uri=track.url)
|
||||||
|
|
||||||
@@ -533,13 +527,9 @@ class HLS:
|
|||||||
if isinstance(res, requests.Response):
|
if isinstance(res, requests.Response):
|
||||||
res.raise_for_status()
|
res.raise_for_status()
|
||||||
init_content = res.content
|
init_content = res.content
|
||||||
elif isinstance(res, httpx.Response):
|
|
||||||
if res.status_code >= 400:
|
|
||||||
raise requests.HTTPError(f"HTTP Error: {res.status_code}", response=res)
|
|
||||||
init_content = res.content
|
|
||||||
else:
|
else:
|
||||||
raise TypeError(
|
raise TypeError(
|
||||||
f"Expected response to be requests.Response or httpx.Response, not {type(res)}"
|
f"Expected response to be requests.Response or curl_cffi.Response, not {type(res)}"
|
||||||
)
|
)
|
||||||
|
|
||||||
map_data = (segment.init_section, init_content)
|
map_data = (segment.init_section, init_content)
|
||||||
@@ -707,7 +697,7 @@ class HLS:
|
|||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def parse_session_data_keys(
|
def parse_session_data_keys(
|
||||||
manifest: M3U8, session: Optional[Union[Session, httpx.Client]] = None
|
manifest: M3U8, session: Optional[Union[Session, CurlSession]] = None
|
||||||
) -> list[m3u8.model.Key]:
|
) -> list[m3u8.model.Key]:
|
||||||
"""Parse `com.apple.hls.keys` session data and return Key objects."""
|
"""Parse `com.apple.hls.keys` session data and return Key objects."""
|
||||||
keys: list[m3u8.model.Key] = []
|
keys: list[m3u8.model.Key] = []
|
||||||
@@ -798,7 +788,8 @@ class HLS:
|
|||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def get_drm(
|
def get_drm(
|
||||||
key: Union[m3u8.model.SessionKey, m3u8.model.Key], session: Optional[Union[Session, httpx.Client]] = None
|
key: Union[m3u8.model.SessionKey, m3u8.model.Key],
|
||||||
|
session: Optional[Union[Session, CurlSession]] = None,
|
||||||
) -> DRM_T:
|
) -> DRM_T:
|
||||||
"""
|
"""
|
||||||
Convert HLS EXT-X-KEY data to an initialized DRM object.
|
Convert HLS EXT-X-KEY data to an initialized DRM object.
|
||||||
@@ -810,8 +801,8 @@ class HLS:
|
|||||||
|
|
||||||
Raises a NotImplementedError if the key system is not supported.
|
Raises a NotImplementedError if the key system is not supported.
|
||||||
"""
|
"""
|
||||||
if not isinstance(session, (Session, httpx.Client, type(None))):
|
if not isinstance(session, (Session, CurlSession, type(None))):
|
||||||
raise TypeError(f"Expected session to be a {Session} or {httpx.Client}, not {type(session)}")
|
raise TypeError(f"Expected session to be a {Session} or {CurlSession}, not {type(session)}")
|
||||||
if not session:
|
if not session:
|
||||||
session = Session()
|
session = Session()
|
||||||
|
|
||||||
|
|||||||
@@ -10,6 +10,7 @@ from pathlib import Path
|
|||||||
from typing import Any, Callable, Optional, Union
|
from typing import Any, Callable, Optional, Union
|
||||||
|
|
||||||
import requests
|
import requests
|
||||||
|
from curl_cffi.requests import Session as CurlSession
|
||||||
from langcodes import Language, tag_is_valid
|
from langcodes import Language, tag_is_valid
|
||||||
from lxml.etree import Element
|
from lxml.etree import Element
|
||||||
from pyplayready.system.pssh import PSSH as PR_PSSH
|
from pyplayready.system.pssh import PSSH as PR_PSSH
|
||||||
@@ -34,11 +35,13 @@ class ISM:
|
|||||||
self.url = url
|
self.url = url
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def from_url(cls, url: str, session: Optional[Session] = None, **kwargs: Any) -> "ISM":
|
def from_url(cls, url: str, session: Optional[Union[Session, CurlSession]] = None, **kwargs: Any) -> "ISM":
|
||||||
if not url:
|
if not url:
|
||||||
raise requests.URLRequired("ISM manifest URL must be provided")
|
raise requests.URLRequired("ISM manifest URL must be provided")
|
||||||
if not session:
|
if not session:
|
||||||
session = Session()
|
session = Session()
|
||||||
|
elif not isinstance(session, (Session, CurlSession)):
|
||||||
|
raise TypeError(f"Expected session to be a {Session} or {CurlSession}, not {session!r}")
|
||||||
res = session.get(url, **kwargs)
|
res = session.get(url, **kwargs)
|
||||||
if res.url != url:
|
if res.url != url:
|
||||||
url = res.url
|
url = res.url
|
||||||
|
|||||||
@@ -4,15 +4,10 @@ from __future__ import annotations
|
|||||||
|
|
||||||
from typing import Optional, Union
|
from typing import Optional, Union
|
||||||
|
|
||||||
import httpx
|
|
||||||
import m3u8
|
import m3u8
|
||||||
from pyplayready.cdm import Cdm as PlayReadyCdm
|
from curl_cffi.requests import Session as CurlSession
|
||||||
from pyplayready.system.pssh import PSSH as PR_PSSH
|
|
||||||
from pywidevine.cdm import Cdm as WidevineCdm
|
|
||||||
from pywidevine.pssh import PSSH as WV_PSSH
|
|
||||||
from requests import Session
|
from requests import Session
|
||||||
|
|
||||||
from unshackle.core.drm import PlayReady, Widevine
|
|
||||||
from unshackle.core.manifests.hls import HLS
|
from unshackle.core.manifests.hls import HLS
|
||||||
from unshackle.core.tracks import Tracks
|
from unshackle.core.tracks import Tracks
|
||||||
|
|
||||||
@@ -21,54 +16,17 @@ def parse(
|
|||||||
master: m3u8.M3U8,
|
master: m3u8.M3U8,
|
||||||
language: str,
|
language: str,
|
||||||
*,
|
*,
|
||||||
session: Optional[Union[Session, httpx.Client]] = None,
|
session: Optional[Union[Session, CurlSession]] = None,
|
||||||
) -> Tracks:
|
) -> Tracks:
|
||||||
"""Parse a variant playlist to ``Tracks`` with DRM information."""
|
"""Parse a variant playlist to ``Tracks`` with basic information, defer DRM loading."""
|
||||||
tracks = HLS(master, session=session).to_tracks(language)
|
tracks = HLS(master, session=session).to_tracks(language)
|
||||||
|
|
||||||
need_wv = not any(isinstance(d, Widevine) for t in tracks for d in (t.drm or []))
|
bool(master.session_keys or HLS.parse_session_data_keys(master, session or Session()))
|
||||||
need_pr = not any(isinstance(d, PlayReady) for t in tracks for d in (t.drm or []))
|
|
||||||
|
|
||||||
if (need_wv or need_pr) and tracks.videos:
|
if True:
|
||||||
if not session:
|
for t in tracks.videos + tracks.audio:
|
||||||
session = Session()
|
t.needs_drm_loading = True
|
||||||
|
t.session = session
|
||||||
session_keys = list(master.session_keys or [])
|
|
||||||
session_keys.extend(HLS.parse_session_data_keys(master, session))
|
|
||||||
|
|
||||||
for drm_obj in HLS.get_all_drm(session_keys):
|
|
||||||
if need_wv and isinstance(drm_obj, Widevine):
|
|
||||||
for t in tracks.videos + tracks.audio:
|
|
||||||
t.drm = [d for d in (t.drm or []) if not isinstance(d, Widevine)] + [drm_obj]
|
|
||||||
need_wv = False
|
|
||||||
elif need_pr and isinstance(drm_obj, PlayReady):
|
|
||||||
for t in tracks.videos + tracks.audio:
|
|
||||||
t.drm = [d for d in (t.drm or []) if not isinstance(d, PlayReady)] + [drm_obj]
|
|
||||||
need_pr = False
|
|
||||||
if not need_wv and not need_pr:
|
|
||||||
break
|
|
||||||
|
|
||||||
if (need_wv or need_pr) and tracks.videos:
|
|
||||||
first_video = tracks.videos[0]
|
|
||||||
playlist = m3u8.load(first_video.url)
|
|
||||||
for key in playlist.keys or []:
|
|
||||||
if not key or not key.keyformat:
|
|
||||||
continue
|
|
||||||
fmt = key.keyformat.lower()
|
|
||||||
if need_wv and fmt == WidevineCdm.urn:
|
|
||||||
pssh_b64 = key.uri.split(",")[-1]
|
|
||||||
drm = Widevine(pssh=WV_PSSH(pssh_b64))
|
|
||||||
for t in tracks.videos + tracks.audio:
|
|
||||||
t.drm = [d for d in (t.drm or []) if not isinstance(d, Widevine)] + [drm]
|
|
||||||
need_wv = False
|
|
||||||
elif need_pr and (fmt == PlayReadyCdm or "com.microsoft.playready" in fmt):
|
|
||||||
pssh_b64 = key.uri.split(",")[-1]
|
|
||||||
drm = PlayReady(pssh=PR_PSSH(pssh_b64), pssh_b64=pssh_b64)
|
|
||||||
for t in tracks.videos + tracks.audio:
|
|
||||||
t.drm = [d for d in (t.drm or []) if not isinstance(d, PlayReady)] + [drm]
|
|
||||||
need_pr = False
|
|
||||||
if not need_wv and not need_pr:
|
|
||||||
break
|
|
||||||
|
|
||||||
return tracks
|
return tracks
|
||||||
|
|
||||||
|
|||||||
79
unshackle/core/session.py
Normal file
79
unshackle/core/session.py
Normal file
@@ -0,0 +1,79 @@
|
|||||||
|
"""Session utilities for creating HTTP sessions with different backends."""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
import warnings
|
||||||
|
|
||||||
|
from curl_cffi.requests import Session as CurlSession
|
||||||
|
|
||||||
|
from unshackle.core.config import config
|
||||||
|
|
||||||
|
# Globally suppress curl_cffi HTTPS proxy warnings since some proxy providers
|
||||||
|
# (like NordVPN) require HTTPS URLs but curl_cffi expects HTTP format
|
||||||
|
warnings.filterwarnings(
|
||||||
|
"ignore", message="Make sure you are using https over https proxy.*", category=RuntimeWarning, module="curl_cffi.*"
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
class Session(CurlSession):
|
||||||
|
"""curl_cffi Session with warning suppression."""
|
||||||
|
|
||||||
|
def request(self, method, url, **kwargs):
|
||||||
|
with warnings.catch_warnings():
|
||||||
|
warnings.filterwarnings(
|
||||||
|
"ignore", message="Make sure you are using https over https proxy.*", category=RuntimeWarning
|
||||||
|
)
|
||||||
|
return super().request(method, url, **kwargs)
|
||||||
|
|
||||||
|
|
||||||
|
def session(browser: str | None = None, **kwargs) -> Session:
|
||||||
|
"""
|
||||||
|
Create a curl_cffi session that impersonates a browser.
|
||||||
|
|
||||||
|
This is a full replacement for requests.Session with browser impersonation
|
||||||
|
and anti-bot capabilities. The session uses curl-impersonate under the hood
|
||||||
|
to mimic real browser behavior.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
browser: Browser to impersonate (e.g. "chrome124", "firefox", "safari").
|
||||||
|
Uses the configured default from curl_impersonate.browser if not specified.
|
||||||
|
See https://github.com/lexiforest/curl_cffi#sessions for available options.
|
||||||
|
**kwargs: Additional arguments passed to CurlSession constructor:
|
||||||
|
- headers: Additional headers (dict)
|
||||||
|
- cookies: Cookie jar or dict
|
||||||
|
- auth: HTTP basic auth tuple (username, password)
|
||||||
|
- proxies: Proxy configuration dict
|
||||||
|
- verify: SSL certificate verification (bool, default True)
|
||||||
|
- timeout: Request timeout in seconds (float or tuple)
|
||||||
|
- allow_redirects: Follow redirects (bool, default True)
|
||||||
|
- max_redirects: Maximum redirect count (int)
|
||||||
|
- cert: Client certificate (str or tuple)
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
curl_cffi.requests.Session configured with browser impersonation, common headers,
|
||||||
|
and equivalent retry behavior to requests.Session.
|
||||||
|
|
||||||
|
Example:
|
||||||
|
from unshackle.core.session import session
|
||||||
|
|
||||||
|
class MyService(Service):
|
||||||
|
@staticmethod
|
||||||
|
def get_session():
|
||||||
|
return session() # Uses config default browser
|
||||||
|
"""
|
||||||
|
if browser is None:
|
||||||
|
browser = config.curl_impersonate.get("browser", "chrome124")
|
||||||
|
|
||||||
|
session_config = {
|
||||||
|
"impersonate": browser,
|
||||||
|
"timeout": 30.0,
|
||||||
|
"allow_redirects": True,
|
||||||
|
"max_redirects": 15,
|
||||||
|
"verify": True,
|
||||||
|
}
|
||||||
|
|
||||||
|
session_config.update(kwargs)
|
||||||
|
session_obj = Session(**session_config)
|
||||||
|
session_obj.headers.update(config.headers)
|
||||||
|
|
||||||
|
return session_obj
|
||||||
@@ -13,6 +13,7 @@ from typing import Any, Callable, Iterable, Optional, Union
|
|||||||
from uuid import UUID
|
from uuid import UUID
|
||||||
from zlib import crc32
|
from zlib import crc32
|
||||||
|
|
||||||
|
from curl_cffi.requests import Session as CurlSession
|
||||||
from langcodes import Language
|
from langcodes import Language
|
||||||
from pyplayready.cdm import Cdm as PlayReadyCdm
|
from pyplayready.cdm import Cdm as PlayReadyCdm
|
||||||
from pywidevine.cdm import Cdm as WidevineCdm
|
from pywidevine.cdm import Cdm as WidevineCdm
|
||||||
@@ -420,6 +421,15 @@ class Track:
|
|||||||
for drm in self.drm:
|
for drm in self.drm:
|
||||||
if isinstance(drm, PlayReady):
|
if isinstance(drm, PlayReady):
|
||||||
return drm
|
return drm
|
||||||
|
elif hasattr(cdm, "is_playready"):
|
||||||
|
if cdm.is_playready:
|
||||||
|
for drm in self.drm:
|
||||||
|
if isinstance(drm, PlayReady):
|
||||||
|
return drm
|
||||||
|
else:
|
||||||
|
for drm in self.drm:
|
||||||
|
if isinstance(drm, Widevine):
|
||||||
|
return drm
|
||||||
|
|
||||||
return self.drm[0]
|
return self.drm[0]
|
||||||
|
|
||||||
@@ -464,6 +474,83 @@ class Track:
|
|||||||
if tenc.key_ID.int != 0:
|
if tenc.key_ID.int != 0:
|
||||||
return tenc.key_ID
|
return tenc.key_ID
|
||||||
|
|
||||||
|
def load_drm_if_needed(self, service=None) -> bool:
|
||||||
|
"""
|
||||||
|
Load DRM information for this track if it was deferred during parsing.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
service: Service instance that can fetch track-specific DRM info
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
True if DRM was loaded or already present, False if failed
|
||||||
|
"""
|
||||||
|
if not getattr(self, "needs_drm_loading", False):
|
||||||
|
return bool(self.drm)
|
||||||
|
|
||||||
|
if self.drm:
|
||||||
|
self.needs_drm_loading = False
|
||||||
|
return True
|
||||||
|
|
||||||
|
if not service or not hasattr(service, "get_track_drm"):
|
||||||
|
return self.load_drm_from_playlist()
|
||||||
|
|
||||||
|
try:
|
||||||
|
track_drm = service.get_track_drm(self)
|
||||||
|
if track_drm:
|
||||||
|
self.drm = track_drm if isinstance(track_drm, list) else [track_drm]
|
||||||
|
self.needs_drm_loading = False
|
||||||
|
return True
|
||||||
|
except Exception as e:
|
||||||
|
raise ValueError(f"Failed to load DRM from service for track {self.id}: {e}")
|
||||||
|
|
||||||
|
return self.load_drm_from_playlist()
|
||||||
|
|
||||||
|
def load_drm_from_playlist(self) -> bool:
|
||||||
|
"""
|
||||||
|
Fallback method to load DRM by fetching this track's individual playlist.
|
||||||
|
"""
|
||||||
|
if self.drm:
|
||||||
|
self.needs_drm_loading = False
|
||||||
|
return True
|
||||||
|
|
||||||
|
try:
|
||||||
|
import m3u8
|
||||||
|
from pyplayready.cdm import Cdm as PlayReadyCdm
|
||||||
|
from pyplayready.system.pssh import PSSH as PR_PSSH
|
||||||
|
from pywidevine.cdm import Cdm as WidevineCdm
|
||||||
|
from pywidevine.pssh import PSSH as WV_PSSH
|
||||||
|
|
||||||
|
session = getattr(self, "session", None) or Session()
|
||||||
|
|
||||||
|
response = session.get(self.url)
|
||||||
|
playlist = m3u8.loads(response.text, self.url)
|
||||||
|
|
||||||
|
drm_list = []
|
||||||
|
|
||||||
|
for key in playlist.keys or []:
|
||||||
|
if not key or not key.keyformat:
|
||||||
|
continue
|
||||||
|
|
||||||
|
fmt = key.keyformat.lower()
|
||||||
|
if fmt == WidevineCdm.urn:
|
||||||
|
pssh_b64 = key.uri.split(",")[-1]
|
||||||
|
drm = Widevine(pssh=WV_PSSH(pssh_b64))
|
||||||
|
drm_list.append(drm)
|
||||||
|
elif fmt == PlayReadyCdm or "com.microsoft.playready" in fmt:
|
||||||
|
pssh_b64 = key.uri.split(",")[-1]
|
||||||
|
drm = PlayReady(pssh=PR_PSSH(pssh_b64), pssh_b64=pssh_b64)
|
||||||
|
drm_list.append(drm)
|
||||||
|
|
||||||
|
if drm_list:
|
||||||
|
self.drm = drm_list
|
||||||
|
self.needs_drm_loading = False
|
||||||
|
return True
|
||||||
|
|
||||||
|
except Exception as e:
|
||||||
|
raise ValueError(f"Failed to load DRM from playlist for track {self.id}: {e}")
|
||||||
|
|
||||||
|
return False
|
||||||
|
|
||||||
def get_init_segment(
|
def get_init_segment(
|
||||||
self,
|
self,
|
||||||
maximum_size: int = 20000,
|
maximum_size: int = 20000,
|
||||||
@@ -499,8 +586,8 @@ class Track:
|
|||||||
raise TypeError(f"Expected url to be a {str}, not {type(url)}")
|
raise TypeError(f"Expected url to be a {str}, not {type(url)}")
|
||||||
if not isinstance(byte_range, (str, type(None))):
|
if not isinstance(byte_range, (str, type(None))):
|
||||||
raise TypeError(f"Expected byte_range to be a {str}, not {type(byte_range)}")
|
raise TypeError(f"Expected byte_range to be a {str}, not {type(byte_range)}")
|
||||||
if not isinstance(session, (Session, type(None))):
|
if not isinstance(session, (Session, CurlSession, type(None))):
|
||||||
raise TypeError(f"Expected session to be a {Session}, not {type(session)}")
|
raise TypeError(f"Expected session to be a {Session} or {CurlSession}, not {type(session)}")
|
||||||
|
|
||||||
if not url:
|
if not url:
|
||||||
if self.descriptor != self.Descriptor.URL:
|
if self.descriptor != self.Descriptor.URL:
|
||||||
@@ -558,15 +645,32 @@ class Track:
|
|||||||
output_path = original_path.with_stem(f"{original_path.stem}_repack")
|
output_path = original_path.with_stem(f"{original_path.stem}_repack")
|
||||||
|
|
||||||
def _ffmpeg(extra_args: list[str] = None):
|
def _ffmpeg(extra_args: list[str] = None):
|
||||||
subprocess.run(
|
args = [
|
||||||
|
binaries.FFMPEG,
|
||||||
|
"-hide_banner",
|
||||||
|
"-loglevel",
|
||||||
|
"error",
|
||||||
|
"-i",
|
||||||
|
original_path,
|
||||||
|
*(extra_args or []),
|
||||||
|
]
|
||||||
|
|
||||||
|
if hasattr(self, "data") and self.data.get("audio_language"):
|
||||||
|
audio_lang = self.data["audio_language"]
|
||||||
|
audio_name = self.data.get("audio_language_name", audio_lang)
|
||||||
|
args.extend(
|
||||||
|
[
|
||||||
|
"-metadata:s:a:0",
|
||||||
|
f"language={audio_lang}",
|
||||||
|
"-metadata:s:a:0",
|
||||||
|
f"title={audio_name}",
|
||||||
|
"-metadata:s:a:0",
|
||||||
|
f"handler_name={audio_name}",
|
||||||
|
]
|
||||||
|
)
|
||||||
|
|
||||||
|
args.extend(
|
||||||
[
|
[
|
||||||
binaries.FFMPEG,
|
|
||||||
"-hide_banner",
|
|
||||||
"-loglevel",
|
|
||||||
"error",
|
|
||||||
"-i",
|
|
||||||
original_path,
|
|
||||||
*(extra_args or []),
|
|
||||||
# Following are very important!
|
# Following are very important!
|
||||||
"-map_metadata",
|
"-map_metadata",
|
||||||
"-1", # don't transfer metadata to output file
|
"-1", # don't transfer metadata to output file
|
||||||
@@ -575,7 +679,11 @@ class Track:
|
|||||||
"-codec",
|
"-codec",
|
||||||
"copy",
|
"copy",
|
||||||
str(output_path),
|
str(output_path),
|
||||||
],
|
]
|
||||||
|
)
|
||||||
|
|
||||||
|
subprocess.run(
|
||||||
|
args,
|
||||||
check=True,
|
check=True,
|
||||||
stdout=subprocess.PIPE,
|
stdout=subprocess.PIPE,
|
||||||
stderr=subprocess.PIPE,
|
stderr=subprocess.PIPE,
|
||||||
|
|||||||
@@ -181,7 +181,7 @@ class Tracks:
|
|||||||
log = logging.getLogger("Tracks")
|
log = logging.getLogger("Tracks")
|
||||||
|
|
||||||
if duplicates:
|
if duplicates:
|
||||||
log.warning(f" - Found and skipped {duplicates} duplicate tracks...")
|
log.debug(f" - Found and skipped {duplicates} duplicate tracks...")
|
||||||
|
|
||||||
def sort_videos(self, by_language: Optional[Sequence[Union[str, Language]]] = None) -> None:
|
def sort_videos(self, by_language: Optional[Sequence[Union[str, Language]]] = None) -> None:
|
||||||
"""Sort video tracks by bitrate, and optionally language."""
|
"""Sort video tracks by bitrate, and optionally language."""
|
||||||
@@ -305,7 +305,14 @@ class Tracks:
|
|||||||
)
|
)
|
||||||
return selected
|
return selected
|
||||||
|
|
||||||
def mux(self, title: str, delete: bool = True, progress: Optional[partial] = None) -> tuple[Path, int, list[str]]:
|
def mux(
|
||||||
|
self,
|
||||||
|
title: str,
|
||||||
|
delete: bool = True,
|
||||||
|
progress: Optional[partial] = None,
|
||||||
|
audio_expected: bool = True,
|
||||||
|
title_language: Optional[Language] = None,
|
||||||
|
) -> tuple[Path, int, list[str]]:
|
||||||
"""
|
"""
|
||||||
Multiplex all the Tracks into a Matroska Container file.
|
Multiplex all the Tracks into a Matroska Container file.
|
||||||
|
|
||||||
@@ -315,7 +322,28 @@ class Tracks:
|
|||||||
delete: Delete all track files after multiplexing.
|
delete: Delete all track files after multiplexing.
|
||||||
progress: Update a rich progress bar via `completed=...`. This must be the
|
progress: Update a rich progress bar via `completed=...`. This must be the
|
||||||
progress object's update() func, pre-set with task id via functools.partial.
|
progress object's update() func, pre-set with task id via functools.partial.
|
||||||
|
audio_expected: Whether audio is expected in the output. Used to determine
|
||||||
|
if embedded audio metadata should be added.
|
||||||
|
title_language: The title's intended language. Used to select the best video track
|
||||||
|
for audio metadata when multiple video tracks exist.
|
||||||
"""
|
"""
|
||||||
|
if self.videos and not self.audio and audio_expected:
|
||||||
|
video_track = None
|
||||||
|
if title_language:
|
||||||
|
video_track = next((v for v in self.videos if v.language == title_language), None)
|
||||||
|
if not video_track:
|
||||||
|
video_track = next((v for v in self.videos if v.is_original_lang), None)
|
||||||
|
|
||||||
|
video_track = video_track or self.videos[0]
|
||||||
|
if video_track.language.is_valid():
|
||||||
|
lang_code = str(video_track.language)
|
||||||
|
lang_name = video_track.language.display_name()
|
||||||
|
|
||||||
|
for video in self.videos:
|
||||||
|
video.needs_repack = True
|
||||||
|
video.data["audio_language"] = lang_code
|
||||||
|
video.data["audio_language_name"] = lang_name
|
||||||
|
|
||||||
if not binaries.MKVToolNix:
|
if not binaries.MKVToolNix:
|
||||||
raise RuntimeError("MKVToolNix (mkvmerge) is required for muxing but was not found")
|
raise RuntimeError("MKVToolNix (mkvmerge) is required for muxing but was not found")
|
||||||
|
|
||||||
@@ -332,12 +360,20 @@ class Tracks:
|
|||||||
raise ValueError("Video Track must be downloaded before muxing...")
|
raise ValueError("Video Track must be downloaded before muxing...")
|
||||||
events.emit(events.Types.TRACK_MULTIPLEX, track=vt)
|
events.emit(events.Types.TRACK_MULTIPLEX, track=vt)
|
||||||
|
|
||||||
|
is_default = False
|
||||||
|
if title_language:
|
||||||
|
is_default = vt.language == title_language
|
||||||
|
if not any(v.language == title_language for v in self.videos):
|
||||||
|
is_default = vt.is_original_lang or i == 0
|
||||||
|
else:
|
||||||
|
is_default = i == 0
|
||||||
|
|
||||||
# Prepare base arguments
|
# Prepare base arguments
|
||||||
video_args = [
|
video_args = [
|
||||||
"--language",
|
"--language",
|
||||||
f"0:{vt.language}",
|
f"0:{vt.language}",
|
||||||
"--default-track",
|
"--default-track",
|
||||||
f"0:{i == 0}",
|
f"0:{is_default}",
|
||||||
"--original-flag",
|
"--original-flag",
|
||||||
f"0:{vt.is_original_lang}",
|
f"0:{vt.is_original_lang}",
|
||||||
"--compression",
|
"--compression",
|
||||||
@@ -363,6 +399,18 @@ class Tracks:
|
|||||||
]
|
]
|
||||||
)
|
)
|
||||||
|
|
||||||
|
if hasattr(vt, "data") and vt.data.get("audio_language"):
|
||||||
|
audio_lang = vt.data["audio_language"]
|
||||||
|
audio_name = vt.data.get("audio_language_name", audio_lang)
|
||||||
|
video_args.extend(
|
||||||
|
[
|
||||||
|
"--language",
|
||||||
|
f"1:{audio_lang}",
|
||||||
|
"--track-name",
|
||||||
|
f"1:{audio_name}",
|
||||||
|
]
|
||||||
|
)
|
||||||
|
|
||||||
cl.extend(video_args + ["(", str(vt.path), ")"])
|
cl.extend(video_args + ["(", str(vt.path), ")"])
|
||||||
|
|
||||||
for i, at in enumerate(self.audio):
|
for i, at in enumerate(self.audio):
|
||||||
|
|||||||
@@ -8,6 +8,7 @@ import tempfile
|
|||||||
from difflib import SequenceMatcher
|
from difflib import SequenceMatcher
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
from typing import Optional, Tuple
|
from typing import Optional, Tuple
|
||||||
|
from xml.sax.saxutils import escape
|
||||||
|
|
||||||
import requests
|
import requests
|
||||||
from requests.adapters import HTTPAdapter, Retry
|
from requests.adapters import HTTPAdapter, Retry
|
||||||
@@ -289,9 +290,9 @@ def _apply_tags(path: Path, tags: dict[str, str]) -> None:
|
|||||||
log.debug("mkvpropedit not found on PATH; skipping tags")
|
log.debug("mkvpropedit not found on PATH; skipping tags")
|
||||||
return
|
return
|
||||||
log.debug("Applying tags to %s: %s", path, tags)
|
log.debug("Applying tags to %s: %s", path, tags)
|
||||||
xml_lines = ["<?xml version='1.0' encoding='UTF-8'?>", "<Tags>", " <Tag>", " <Targets/>"]
|
xml_lines = ['<?xml version="1.0" encoding="UTF-8"?>', "<Tags>", " <Tag>", " <Targets/>"]
|
||||||
for name, value in tags.items():
|
for name, value in tags.items():
|
||||||
xml_lines.append(f" <Simple><Name>{name}</Name><String>{value}</String></Simple>")
|
xml_lines.append(f" <Simple><Name>{escape(name)}</Name><String>{escape(value)}</String></Simple>")
|
||||||
xml_lines.extend([" </Tag>", "</Tags>"])
|
xml_lines.extend([" </Tag>", "</Tags>"])
|
||||||
with tempfile.NamedTemporaryFile("w", suffix=".xml", delete=False) as f:
|
with tempfile.NamedTemporaryFile("w", suffix=".xml", delete=False) as f:
|
||||||
f.write("\n".join(xml_lines))
|
f.write("\n".join(xml_lines))
|
||||||
@@ -349,13 +350,25 @@ def tag_file(path: Path, title: Title, tmdb_id: Optional[int] | None = None) ->
|
|||||||
if simkl_tmdb_id:
|
if simkl_tmdb_id:
|
||||||
tmdb_id = simkl_tmdb_id
|
tmdb_id = simkl_tmdb_id
|
||||||
|
|
||||||
show_ids = simkl_data.get("show", {}).get("ids", {})
|
# Handle TV show data from Simkl
|
||||||
if show_ids.get("imdb"):
|
if simkl_data.get("type") == "episode" and "show" in simkl_data:
|
||||||
standard_tags["IMDB"] = f"https://www.imdb.com/title/{show_ids['imdb']}"
|
show_ids = simkl_data.get("show", {}).get("ids", {})
|
||||||
if show_ids.get("tvdb"):
|
if show_ids.get("imdb"):
|
||||||
standard_tags["TVDB"] = f"https://thetvdb.com/dereferrer/series/{show_ids['tvdb']}"
|
standard_tags["IMDB"] = show_ids["imdb"]
|
||||||
if show_ids.get("tmdbtv"):
|
if show_ids.get("tvdb"):
|
||||||
standard_tags["TMDB"] = f"https://www.themoviedb.org/tv/{show_ids['tmdbtv']}"
|
standard_tags["TVDB2"] = f"series/{show_ids['tvdb']}"
|
||||||
|
if show_ids.get("tmdbtv"):
|
||||||
|
standard_tags["TMDB"] = f"tv/{show_ids['tmdbtv']}"
|
||||||
|
|
||||||
|
# Handle movie data from Simkl
|
||||||
|
elif simkl_data.get("type") == "movie" and "movie" in simkl_data:
|
||||||
|
movie_ids = simkl_data.get("movie", {}).get("ids", {})
|
||||||
|
if movie_ids.get("imdb"):
|
||||||
|
standard_tags["IMDB"] = movie_ids["imdb"]
|
||||||
|
if movie_ids.get("tvdb"):
|
||||||
|
standard_tags["TVDB2"] = f"movies/{movie_ids['tvdb']}"
|
||||||
|
if movie_ids.get("tmdb"):
|
||||||
|
standard_tags["TMDB"] = f"movie/{movie_ids['tmdb']}"
|
||||||
|
|
||||||
# Use TMDB API for additional metadata (either from provided ID or Simkl lookup)
|
# Use TMDB API for additional metadata (either from provided ID or Simkl lookup)
|
||||||
api_key = _api_key()
|
api_key = _api_key()
|
||||||
@@ -373,8 +386,8 @@ def tag_file(path: Path, title: Title, tmdb_id: Optional[int] | None = None) ->
|
|||||||
_apply_tags(path, custom_tags)
|
_apply_tags(path, custom_tags)
|
||||||
return
|
return
|
||||||
|
|
||||||
tmdb_url = f"https://www.themoviedb.org/{'movie' if kind == 'movie' else 'tv'}/{tmdb_id}"
|
prefix = "movie" if kind == "movie" else "tv"
|
||||||
standard_tags["TMDB"] = tmdb_url
|
standard_tags["TMDB"] = f"{prefix}/{tmdb_id}"
|
||||||
try:
|
try:
|
||||||
ids = external_ids(tmdb_id, kind)
|
ids = external_ids(tmdb_id, kind)
|
||||||
except requests.RequestException as exc:
|
except requests.RequestException as exc:
|
||||||
@@ -385,11 +398,13 @@ def tag_file(path: Path, title: Title, tmdb_id: Optional[int] | None = None) ->
|
|||||||
|
|
||||||
imdb_id = ids.get("imdb_id")
|
imdb_id = ids.get("imdb_id")
|
||||||
if imdb_id:
|
if imdb_id:
|
||||||
standard_tags["IMDB"] = f"https://www.imdb.com/title/{imdb_id}"
|
standard_tags["IMDB"] = imdb_id
|
||||||
tvdb_id = ids.get("tvdb_id")
|
tvdb_id = ids.get("tvdb_id")
|
||||||
if tvdb_id:
|
if tvdb_id:
|
||||||
tvdb_prefix = "movies" if kind == "movie" else "series"
|
if kind == "movie":
|
||||||
standard_tags["TVDB"] = f"https://thetvdb.com/dereferrer/{tvdb_prefix}/{tvdb_id}"
|
standard_tags["TVDB2"] = f"movies/{tvdb_id}"
|
||||||
|
else:
|
||||||
|
standard_tags["TVDB2"] = f"series/{tvdb_id}"
|
||||||
|
|
||||||
merged_tags = {
|
merged_tags = {
|
||||||
**custom_tags,
|
**custom_tags,
|
||||||
|
|||||||
@@ -74,7 +74,9 @@ class Vaults:
|
|||||||
for vault in self.vaults:
|
for vault in self.vaults:
|
||||||
if not vault.no_push:
|
if not vault.no_push:
|
||||||
try:
|
try:
|
||||||
success += bool(vault.add_keys(self.service, kid_keys))
|
# Count each vault that successfully processes the keys (whether new or existing)
|
||||||
|
vault.add_keys(self.service, kid_keys)
|
||||||
|
success += 1
|
||||||
except (PermissionError, NotImplementedError):
|
except (PermissionError, NotImplementedError):
|
||||||
pass
|
pass
|
||||||
return success
|
return success
|
||||||
|
|||||||
@@ -282,6 +282,10 @@ class EXAMPLE(Service):
|
|||||||
|
|
||||||
return chapters
|
return chapters
|
||||||
|
|
||||||
|
def get_widevine_service_certificate(self, **_: any) -> str:
|
||||||
|
"""Return the Widevine service certificate from config, if available."""
|
||||||
|
return self.config.get("certificate")
|
||||||
|
|
||||||
def get_playready_license(self, *, challenge: bytes, title: Title_T, track: AnyTrack) -> Optional[bytes]:
|
def get_playready_license(self, *, challenge: bytes, title: Title_T, track: AnyTrack) -> Optional[bytes]:
|
||||||
"""Retrieve a PlayReady license for a given track."""
|
"""Retrieve a PlayReady license for a given track."""
|
||||||
|
|
||||||
|
|||||||
@@ -88,6 +88,26 @@ cdm:
|
|||||||
jane_uhd: nexus_5_l1 # Profile 'jane_uhd' uses Nexus 5 L1
|
jane_uhd: nexus_5_l1 # Profile 'jane_uhd' uses Nexus 5 L1
|
||||||
default: generic_android_l3 # Default CDM for this service
|
default: generic_android_l3 # Default CDM for this service
|
||||||
|
|
||||||
|
# NEW: Quality-based CDM selection
|
||||||
|
# Use different CDMs based on video resolution
|
||||||
|
# Supports operators: >=, >, <=, <, or exact match
|
||||||
|
EXAMPLE_QUALITY:
|
||||||
|
"<=1080": generic_android_l3 # Use L3 for 1080p and below
|
||||||
|
">1080": nexus_5_l1 # Use L1 for above 1080p (1440p, 2160p)
|
||||||
|
default: generic_android_l3 # Optional: fallback if no quality match
|
||||||
|
|
||||||
|
# You can mix profiles and quality thresholds in the same service
|
||||||
|
NETFLIX:
|
||||||
|
# Profile-based selection (existing functionality)
|
||||||
|
john: netflix_l3_profile
|
||||||
|
jane: netflix_l1_profile
|
||||||
|
# Quality-based selection (new functionality)
|
||||||
|
"<=720": netflix_mobile_l3
|
||||||
|
"1080": netflix_standard_l3
|
||||||
|
">=1440": netflix_premium_l1
|
||||||
|
# Fallback
|
||||||
|
default: netflix_standard_l3
|
||||||
|
|
||||||
# Use pywidevine Serve-compliant Remote CDMs
|
# Use pywidevine Serve-compliant Remote CDMs
|
||||||
remote_cdm:
|
remote_cdm:
|
||||||
- name: "chrome"
|
- name: "chrome"
|
||||||
@@ -105,6 +125,50 @@ remote_cdm:
|
|||||||
host: https://domain-2.com/api
|
host: https://domain-2.com/api
|
||||||
secret: secret_key
|
secret: secret_key
|
||||||
|
|
||||||
|
- name: "decrypt_labs_chrome"
|
||||||
|
type: "decrypt_labs" # Required to identify as DecryptLabs CDM
|
||||||
|
device_name: "ChromeCDM" # Scheme identifier - must match exactly
|
||||||
|
device_type: CHROME
|
||||||
|
system_id: 4464 # Doesn't matter
|
||||||
|
security_level: 3
|
||||||
|
host: "https://keyxtractor.decryptlabs.com"
|
||||||
|
secret: "your_decrypt_labs_api_key_here" # Replace with your API key
|
||||||
|
- name: "decrypt_labs_l1"
|
||||||
|
type: "decrypt_labs"
|
||||||
|
device_name: "L1" # Scheme identifier - must match exactly
|
||||||
|
device_type: ANDROID
|
||||||
|
system_id: 4464
|
||||||
|
security_level: 1
|
||||||
|
host: "https://keyxtractor.decryptlabs.com"
|
||||||
|
secret: "your_decrypt_labs_api_key_here"
|
||||||
|
|
||||||
|
- name: "decrypt_labs_l2"
|
||||||
|
type: "decrypt_labs"
|
||||||
|
device_name: "L2" # Scheme identifier - must match exactly
|
||||||
|
device_type: ANDROID
|
||||||
|
system_id: 4464
|
||||||
|
security_level: 2
|
||||||
|
host: "https://keyxtractor.decryptlabs.com"
|
||||||
|
secret: "your_decrypt_labs_api_key_here"
|
||||||
|
|
||||||
|
- name: "decrypt_labs_playready_sl2"
|
||||||
|
type: "decrypt_labs"
|
||||||
|
device_name: "SL2" # Scheme identifier - must match exactly
|
||||||
|
device_type: PLAYREADY
|
||||||
|
system_id: 0
|
||||||
|
security_level: 2000
|
||||||
|
host: "https://keyxtractor.decryptlabs.com"
|
||||||
|
secret: "your_decrypt_labs_api_key_here"
|
||||||
|
|
||||||
|
- name: "decrypt_labs_playready_sl3"
|
||||||
|
type: "decrypt_labs"
|
||||||
|
device_name: "SL3" # Scheme identifier - must match exactly
|
||||||
|
device_type: PLAYREADY
|
||||||
|
system_id: 0
|
||||||
|
security_level: 3000
|
||||||
|
host: "https://keyxtractor.decryptlabs.com"
|
||||||
|
secret: "your_decrypt_labs_api_key_here"
|
||||||
|
|
||||||
# Key Vaults store your obtained Content Encryption Keys (CEKs)
|
# Key Vaults store your obtained Content Encryption Keys (CEKs)
|
||||||
# Use 'no_push: true' to prevent a vault from receiving pushed keys
|
# Use 'no_push: true' to prevent a vault from receiving pushed keys
|
||||||
# while still allowing it to provide keys when requested
|
# while still allowing it to provide keys when requested
|
||||||
@@ -171,7 +235,7 @@ chapter_fallback_name: "Chapter {j:02}"
|
|||||||
# Case-Insensitive dictionary of headers for all Services
|
# Case-Insensitive dictionary of headers for all Services
|
||||||
headers:
|
headers:
|
||||||
Accept-Language: "en-US,en;q=0.8"
|
Accept-Language: "en-US,en;q=0.8"
|
||||||
User-Agent: "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/77.0.3865.75 Safari/537.36"
|
User-Agent: "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/138.0.0.0 Safari/537.36"
|
||||||
|
|
||||||
# Override default filenames used across unshackle
|
# Override default filenames used across unshackle
|
||||||
filenames:
|
filenames:
|
||||||
@@ -213,6 +277,13 @@ services:
|
|||||||
# Global service config
|
# Global service config
|
||||||
api_key: "service_api_key"
|
api_key: "service_api_key"
|
||||||
|
|
||||||
|
# Service certificate for Widevine L1/L2 (base64 encoded)
|
||||||
|
# This certificate is automatically used when L1/L2 schemes are selected
|
||||||
|
# Services obtain this from their DRM provider or license server
|
||||||
|
certificate: |
|
||||||
|
CAUSwwUKvQIIAxIQ5US6QAvBDzfTtjb4tU/7QxiH8c+TBSKOAjCCAQoCggEBAObzvlu2hZRsapAPx4Aa4GUZj4/GjxgXUtBH4THSkM40x63wQeyVxlEEo
|
||||||
|
# ... (full base64 certificate here)
|
||||||
|
|
||||||
# Profile-specific device configurations
|
# Profile-specific device configurations
|
||||||
profiles:
|
profiles:
|
||||||
john_sd:
|
john_sd:
|
||||||
|
|||||||
@@ -28,26 +28,33 @@ class MySQL(Vault):
|
|||||||
raise PermissionError(f"MySQL vault {self.slug} has no SELECT permission.")
|
raise PermissionError(f"MySQL vault {self.slug} has no SELECT permission.")
|
||||||
|
|
||||||
def get_key(self, kid: Union[UUID, str], service: str) -> Optional[str]:
|
def get_key(self, kid: Union[UUID, str], service: str) -> Optional[str]:
|
||||||
if not self.has_table(service):
|
|
||||||
# no table, no key, simple
|
|
||||||
return None
|
|
||||||
|
|
||||||
if isinstance(kid, UUID):
|
if isinstance(kid, UUID):
|
||||||
kid = kid.hex
|
kid = kid.hex
|
||||||
|
|
||||||
|
service_variants = [service]
|
||||||
|
if service != service.lower():
|
||||||
|
service_variants.append(service.lower())
|
||||||
|
if service != service.upper():
|
||||||
|
service_variants.append(service.upper())
|
||||||
|
|
||||||
conn = self.conn_factory.get()
|
conn = self.conn_factory.get()
|
||||||
cursor = conn.cursor()
|
cursor = conn.cursor()
|
||||||
|
|
||||||
try:
|
try:
|
||||||
cursor.execute(
|
for service_name in service_variants:
|
||||||
# TODO: SQL injection risk
|
if not self.has_table(service_name):
|
||||||
f"SELECT `id`, `key_` FROM `{service}` WHERE `kid`=%s AND `key_`!=%s",
|
continue
|
||||||
(kid, "0" * 32),
|
|
||||||
)
|
cursor.execute(
|
||||||
cek = cursor.fetchone()
|
# TODO: SQL injection risk
|
||||||
if not cek:
|
f"SELECT `id`, `key_` FROM `{service_name}` WHERE `kid`=%s AND `key_`!=%s",
|
||||||
return None
|
(kid, "0" * 32),
|
||||||
return cek["key_"]
|
)
|
||||||
|
cek = cursor.fetchone()
|
||||||
|
if cek:
|
||||||
|
return cek["key_"]
|
||||||
|
|
||||||
|
return None
|
||||||
finally:
|
finally:
|
||||||
cursor.close()
|
cursor.close()
|
||||||
|
|
||||||
@@ -131,16 +138,27 @@ class MySQL(Vault):
|
|||||||
if any(isinstance(kid, UUID) for kid, key_ in kid_keys.items()):
|
if any(isinstance(kid, UUID) for kid, key_ in kid_keys.items()):
|
||||||
kid_keys = {kid.hex if isinstance(kid, UUID) else kid: key_ for kid, key_ in kid_keys.items()}
|
kid_keys = {kid.hex if isinstance(kid, UUID) else kid: key_ for kid, key_ in kid_keys.items()}
|
||||||
|
|
||||||
|
if not kid_keys:
|
||||||
|
return 0
|
||||||
|
|
||||||
conn = self.conn_factory.get()
|
conn = self.conn_factory.get()
|
||||||
cursor = conn.cursor()
|
cursor = conn.cursor()
|
||||||
|
|
||||||
try:
|
try:
|
||||||
|
placeholders = ",".join(["%s"] * len(kid_keys))
|
||||||
|
cursor.execute(f"SELECT kid FROM `{service}` WHERE kid IN ({placeholders})", list(kid_keys.keys()))
|
||||||
|
existing_kids = {row["kid"] for row in cursor.fetchall()}
|
||||||
|
|
||||||
|
new_keys = {kid: key for kid, key in kid_keys.items() if kid not in existing_kids}
|
||||||
|
|
||||||
|
if not new_keys:
|
||||||
|
return 0
|
||||||
|
|
||||||
cursor.executemany(
|
cursor.executemany(
|
||||||
# TODO: SQL injection risk
|
f"INSERT INTO `{service}` (kid, key_) VALUES (%s, %s)",
|
||||||
f"INSERT IGNORE INTO `{service}` (kid, key_) VALUES (%s, %s)",
|
new_keys.items(),
|
||||||
kid_keys.items(),
|
|
||||||
)
|
)
|
||||||
return cursor.rowcount
|
return len(new_keys)
|
||||||
finally:
|
finally:
|
||||||
conn.commit()
|
conn.commit()
|
||||||
cursor.close()
|
cursor.close()
|
||||||
|
|||||||
@@ -19,22 +19,30 @@ class SQLite(Vault):
|
|||||||
self.conn_factory = ConnectionFactory(self.path)
|
self.conn_factory = ConnectionFactory(self.path)
|
||||||
|
|
||||||
def get_key(self, kid: Union[UUID, str], service: str) -> Optional[str]:
|
def get_key(self, kid: Union[UUID, str], service: str) -> Optional[str]:
|
||||||
if not self.has_table(service):
|
|
||||||
# no table, no key, simple
|
|
||||||
return None
|
|
||||||
|
|
||||||
if isinstance(kid, UUID):
|
if isinstance(kid, UUID):
|
||||||
kid = kid.hex
|
kid = kid.hex
|
||||||
|
|
||||||
conn = self.conn_factory.get()
|
conn = self.conn_factory.get()
|
||||||
cursor = conn.cursor()
|
cursor = conn.cursor()
|
||||||
|
|
||||||
|
# Try both the original service name and lowercase version to handle case sensitivity issues
|
||||||
|
service_variants = [service]
|
||||||
|
if service != service.lower():
|
||||||
|
service_variants.append(service.lower())
|
||||||
|
if service != service.upper():
|
||||||
|
service_variants.append(service.upper())
|
||||||
|
|
||||||
try:
|
try:
|
||||||
cursor.execute(f"SELECT `id`, `key_` FROM `{service}` WHERE `kid`=? AND `key_`!=?", (kid, "0" * 32))
|
for service_name in service_variants:
|
||||||
cek = cursor.fetchone()
|
if not self.has_table(service_name):
|
||||||
if not cek:
|
continue
|
||||||
return None
|
|
||||||
return cek[1]
|
cursor.execute(f"SELECT `id`, `key_` FROM `{service_name}` WHERE `kid`=? AND `key_`!=?", (kid, "0" * 32))
|
||||||
|
cek = cursor.fetchone()
|
||||||
|
if cek:
|
||||||
|
return cek[1]
|
||||||
|
|
||||||
|
return None
|
||||||
finally:
|
finally:
|
||||||
cursor.close()
|
cursor.close()
|
||||||
|
|
||||||
@@ -102,16 +110,27 @@ class SQLite(Vault):
|
|||||||
if any(isinstance(kid, UUID) for kid, key_ in kid_keys.items()):
|
if any(isinstance(kid, UUID) for kid, key_ in kid_keys.items()):
|
||||||
kid_keys = {kid.hex if isinstance(kid, UUID) else kid: key_ for kid, key_ in kid_keys.items()}
|
kid_keys = {kid.hex if isinstance(kid, UUID) else kid: key_ for kid, key_ in kid_keys.items()}
|
||||||
|
|
||||||
|
if not kid_keys:
|
||||||
|
return 0
|
||||||
|
|
||||||
conn = self.conn_factory.get()
|
conn = self.conn_factory.get()
|
||||||
cursor = conn.cursor()
|
cursor = conn.cursor()
|
||||||
|
|
||||||
try:
|
try:
|
||||||
|
placeholders = ",".join(["?"] * len(kid_keys))
|
||||||
|
cursor.execute(f"SELECT kid FROM `{service}` WHERE kid IN ({placeholders})", list(kid_keys.keys()))
|
||||||
|
existing_kids = {row[0] for row in cursor.fetchall()}
|
||||||
|
|
||||||
|
new_keys = {kid: key for kid, key in kid_keys.items() if kid not in existing_kids}
|
||||||
|
|
||||||
|
if not new_keys:
|
||||||
|
return 0
|
||||||
|
|
||||||
cursor.executemany(
|
cursor.executemany(
|
||||||
# TODO: SQL injection risk
|
f"INSERT INTO `{service}` (kid, key_) VALUES (?, ?)",
|
||||||
f"INSERT OR IGNORE INTO `{service}` (kid, key_) VALUES (?, ?)",
|
new_keys.items(),
|
||||||
kid_keys.items(),
|
|
||||||
)
|
)
|
||||||
return cursor.rowcount
|
return len(new_keys)
|
||||||
finally:
|
finally:
|
||||||
conn.commit()
|
conn.commit()
|
||||||
cursor.close()
|
cursor.close()
|
||||||
|
|||||||
Reference in New Issue
Block a user