Merge pull request #828 from ouzhuowei/add_save_data_path

补充代理配置的arp
This commit is contained in:
程序员阿江-Relakkes
2026-02-12 04:47:25 +08:00
committed by GitHub

View File

@@ -274,12 +274,38 @@ async def parse_cmd(argv: Optional[Sequence[str]] = None):
rich_help_panel="Storage Configuration",
),
] = config.SAVE_DATA_PATH,
enable_ip_proxy: Annotated[
str,
typer.Option(
"--enable_ip_proxy",
help="Whether to enable IP proxy, supports yes/true/t/y/1 or no/false/f/n/0",
rich_help_panel="Proxy Configuration",
show_default=True,
),
] = str(config.ENABLE_IP_PROXY),
ip_proxy_pool_count: Annotated[
int,
typer.Option(
"--ip_proxy_pool_count",
help="IP proxy pool count",
rich_help_panel="Proxy Configuration",
),
] = config.IP_PROXY_POOL_COUNT,
ip_proxy_provider_name: Annotated[
str,
typer.Option(
"--ip_proxy_provider_name",
help="IP proxy provider name (kuaidaili | wandouhttp)",
rich_help_panel="Proxy Configuration",
),
] = config.IP_PROXY_PROVIDER_NAME,
) -> SimpleNamespace:
"""MediaCrawler 命令行入口"""
enable_comment = _to_bool(get_comment)
enable_sub_comment = _to_bool(get_sub_comment)
enable_headless = _to_bool(headless)
enable_ip_proxy_value = _to_bool(enable_ip_proxy)
init_db_value = init_db.value if init_db else None
# Parse specified_id and creator_id into lists
@@ -301,6 +327,9 @@ async def parse_cmd(argv: Optional[Sequence[str]] = None):
config.CRAWLER_MAX_COMMENTS_COUNT_SINGLENOTES = max_comments_count_singlenotes
config.MAX_CONCURRENCY_NUM = max_concurrency_num
config.SAVE_DATA_PATH = save_data_path
config.ENABLE_IP_PROXY = enable_ip_proxy_value
config.IP_PROXY_POOL_COUNT = ip_proxy_pool_count
config.IP_PROXY_PROVIDER_NAME = ip_proxy_provider_name
# Set platform-specific ID lists for detail/creator mode
if specified_id_list: