-
Notifications
You must be signed in to change notification settings - Fork 16
Expand file tree
/
Copy pathmerge.py
More file actions
276 lines (242 loc) · 10.9 KB
/
merge.py
File metadata and controls
276 lines (242 loc) · 10.9 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
import xml.etree.ElementTree as ET
from collections import defaultdict
import aiohttp
import asyncio
from tqdm.asyncio import tqdm_asyncio # 引入 tqdm 的异步支持
from datetime import datetime, timezone, timedelta
import gzip
import shutil
from xml.dom import minidom
import re
from opencc import OpenCC
import os
from tqdm import tqdm # 引入 tqdm 的同步支持
TZ_UTC_PLUS_8 = timezone(timedelta(hours=8))
# ============ EPG 源预处理规则 ============
# 在合并前对指定源的指定频道做预处理
# 每条规则: (源URL关键字, 频道名关键字, 预处理函数)
def _adjust_timezone(programme, from_offset, to_offset):
"""将 programme 节点的 start/stop 时区从 from_offset 替换为 to_offset"""
for attr in ('start', 'stop'):
val = programme.get(attr, '')
if from_offset in val:
programme.set(attr, val.replace(from_offset, to_offset))
def _make_tz_rule(channel_keyword, from_offset, to_offset):
"""生成一个时区调整规则函数"""
def rule(channel_name, programme):
if channel_keyword in channel_name:
_adjust_timezone(programme, from_offset, to_offset)
return rule
# 预处理规则列表: (源URL包含的关键字, 规则函数)
PREPROCESS_RULES = [
# 天映经典频道: 时区 +0800 → +0900 (延迟一小时)
("kuke31/xmlgz", _make_tz_rule("天映经典", "+0800", "+0700")),
]
def preprocess_epg(url, epg_content):
"""对 epg_content XML 字符串按规则做预处理,返回处理后的字符串"""
matched_rules = [rule for keyword, rule in PREPROCESS_RULES if keyword in url]
if not matched_rules:
return epg_content
try:
parser = ET.XMLParser(encoding='UTF-8')
root = ET.fromstring(epg_content, parser=parser)
except ET.ParseError:
return epg_content
# 建立 channel_id -> display_name 的映射
channel_names = {}
for channel in root.findall('channel'):
cid = channel.get('id', '')
names = [n.text for n in channel.findall('display-name') if n.text]
channel_names[cid] = ' '.join(names) + ' ' + cid
for programme in root.findall('programme'):
cid = programme.get('channel', '')
name_str = channel_names.get(cid, cid)
for rule in matched_rules:
rule(name_str, programme)
return ET.tostring(root, encoding='unicode')
# ============ 预处理规则结束 ============
def transform2_zh_hans(string):
cc = OpenCC("t2s")
new_str = cc.convert(string)
return new_str
async def fetch_epg(url):
connector = aiohttp.TCPConnector(limit=16, ssl=False)
headers = {
"User-Agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/113.0.0.0 Safari/537.36"
}
try:
async with aiohttp.ClientSession(connector=connector, trust_env=True, headers=headers) as session:
async with session.get(url) as response:
if url.endswith('.gz'):
compressed_data = await response.read()
return gzip.decompress(compressed_data).decode('utf-8', errors='ignore')
else:
return await response.text(encoding='utf-8')
except aiohttp.ClientError as e:
print(f"{url}HTTP请求错误: {e}")
except asyncio.TimeoutError:
print("{url}请求超时")
except Exception as e:
print(f"{url}其他错误: {e}")
return None
def process_display_name(display_name):
if display_name.endswith('高清'):
display_name = display_name[:-2]
return display_name
def parse_epg(epg_content):
try:
parser = ET.XMLParser(encoding='UTF-8')
root = ET.fromstring(epg_content, parser=parser)
except ET.ParseError as e:
print(f"Error parsing XML: {e}")
print(f"Problematic content: {epg_content[:500]}")
return {}, defaultdict(list)
channels = {}
programmes = defaultdict(list)
for channel in root.findall('channel'):
channel_id = transform2_zh_hans(channel.get('id'))
channel_display_names = []
for name in channel.findall('display-name'):
t_name = transform2_zh_hans(name.text)
t_name = process_display_name(t_name)
channel_display_names.append([t_name, name.get('lang', 'zh')])
if not channel_id.isdigit() and channel_id not in channel_display_names:
channel_display_names.append([channel_id, 'zh'])
channels[channel_id] = channel_display_names
today = datetime.now(TZ_UTC_PLUS_8).date()
valid_channels = set()
for programme in root.findall('programme'):
channel_id = transform2_zh_hans(programme.get('channel'))
channel_start = datetime.strptime(
re.sub(r'\s+', '', programme.get('start')), "%Y%m%d%H%M%S%z")
channel_stop = datetime.strptime(
re.sub(r'\s+', '', programme.get('stop')), "%Y%m%d%H%M%S%z")
channel_start = channel_start.astimezone(TZ_UTC_PLUS_8)
channel_stop = channel_stop.astimezone(TZ_UTC_PLUS_8)
if channel_stop.date() == today:
valid_channels.add(channel_id)
channel_elem = ET.SubElement(
root, 'programme', attrib={"start": channel_start.strftime("%Y%m%d%H%M%S %z"), "stop": channel_stop.strftime("%Y%m%d%H%M%S %z")})
for title in programme.findall('title'):
if title.text is None:
channel_title = "精彩节目"
else:
channel_title = title.text.strip()
langattr = title.get('lang')
if langattr == 'zh' or langattr is None:
channel_title = transform2_zh_hans(channel_title)
channel_elem_t = ET.SubElement(
channel_elem, 'title')
channel_elem_t.text = channel_title
if langattr is not None:
channel_elem_t.set('lang', langattr)
for desc in programme.findall('desc'):
if desc.text is None:
continue
langattr = desc.get('lang')
channel_desc = desc.text.strip()
if langattr == 'zh' or langattr is None:
channel_desc = transform2_zh_hans(channel_desc)
channel_elem_d = ET.SubElement(
channel_elem, 'desc')
channel_elem_d.text = channel_desc.strip()
if langattr is not None:
channel_elem_d.set('lang', langattr)
programmes[channel_id].append(channel_elem)
# Filter channels that don't have any program ending today
channels = {k: v for k, v in channels.items() if k in valid_channels}
# Optional: Filter programmes as well to keep data consistent,
# though only valid channels are returned so main loop might be fine.
# But filtering programmes dict saves memory and ensures correctness if main iterates programmes keys logic changes.
programmes = {k: v for k, v in programmes.items() if k in valid_channels}
return channels, programmes
def write_to_xml(channels_id, channels_names, programmes, filename):
# 目录不存在
if not os.path.exists('output'):
os.makedirs('output')
current_time = datetime.now(TZ_UTC_PLUS_8).strftime("%Y%m%d%H%M%S %z")
root = ET.Element('tv', attrib={'date': current_time})
for channel_id in channels_id:
channel_elem = ET.SubElement(
root, 'channel', attrib={"id": channel_id})
for display_name_node in channels_names[channel_id]:
display_name = display_name_node[0]
langattr = display_name_node[1]
display_name_elem = ET.SubElement(
channel_elem, 'display-name', attrib={"lang": langattr})
display_name_elem.text = display_name
for prog in programmes[channel_id]:
prog.set('channel', channel_id) # 设置 programme 的 channel 属性
root.append(prog)
# Beautify the XML output
rough_string = ET.tostring(root, 'utf-8')
reparsed = minidom.parseString(rough_string)
with open(filename, 'w', encoding='utf-8') as f:
f.write(reparsed.toprettyxml(indent='\t', newl='\n'))
def compress_to_gz(input_filename, output_filename):
with open(input_filename, 'rb') as f_in:
with gzip.open(output_filename, 'wb') as f_out:
shutil.copyfileobj(f_in, f_out)
def get_urls():
urls = []
with open('config.txt', 'r', encoding='utf-8') as file:
for line in file:
line = line.strip()
if line and not line.startswith('#'):
urls.append(line)
return urls
async def main():
urls = get_urls()
tasks = [fetch_epg(url) for url in urls]
print("Fetching EPG data...")
epg_contents = await tqdm_asyncio.gather(*tasks, desc="Fetching URLs")
all_channels_map = {}
all_channel_id = set()
all_channel_names = defaultdict(list)
all_programmes = defaultdict(list)
print("Finished.")
i = 0
for epg_content in epg_contents:
i += 1
print(f"Processing EPG source...{i}/{len(epg_contents)}")
if epg_content is None:
continue
print("Parsing EPG data...")
epg_content = preprocess_epg(urls[i - 1], epg_content)
channels, programmes = parse_epg(epg_content)
print("Finished.")
with tqdm(total=len(channels), desc="Merging EPG", unit="file") as pbar:
for channel_id, display_names in channels.items():
if len(programmes[channel_id]) == 0:
continue
is_in_map = False
map_id = ""
for display_name_node in display_names:
if is_in_map:
break
display_name = display_name_node[0]
is_in_map = display_name in all_channels_map
map_id = display_name
map_id = all_channels_map.get(map_id, channel_id)
if not is_in_map:
all_channel_id.add(map_id)
all_channel_names[map_id] = display_names
all_programmes[map_id] = programmes[channel_id]
for display_name_node in display_names:
display_name = display_name_node[0]
all_channels_map[display_name] = map_id
else:
if len(all_programmes[map_id]) < len(programmes[channel_id]):
all_programmes[map_id] = programmes[channel_id]
for display_name_node in display_names:
display_name = display_name_node[0]
if display_name not in all_channels_map:
all_channel_names[map_id].append(display_name_node)
all_channels_map[display_name] = map_id
pbar.update(1) # 更新进度条
print("Writing to XML...")
write_to_xml(all_channel_id, all_channel_names,
all_programmes, 'output/epg.xml')
compress_to_gz('output/epg.xml', 'output/epg.gz')
if __name__ == '__main__':
asyncio.run(main())