4
4
from datetime import datetime , timezone
5
5
from django .utils import timezone as django_timezone
6
6
from .models import RSSFeed , RSSItem
7
- import logging
8
-
9
- logger = logging .getLogger (__name__ )
7
+ import logfire
10
8
11
9
12
10
@shared_task
@@ -26,11 +24,11 @@ def crawl_all_rss_feeds():
26
24
result = crawl_single_rss_feed (feed .id )
27
25
results ['processed_feeds' ] += 1
28
26
results ['new_items' ] += result .get ('new_items' , 0 )
29
- logger .info (f"Successfully crawled feed { feed .name } : { result .get ('new_items' , 0 )} new items" )
27
+ logfire .info (f"Successfully crawled feed { feed .name } : { result .get ('new_items' , 0 )} new items" )
30
28
except Exception as e :
31
29
error_msg = f"Error crawling feed { feed .name } : { str (e )} "
32
30
results ['errors' ].append (error_msg )
33
- logger .error (error_msg )
31
+ logfire .error (error_msg )
34
32
35
33
return results
36
34
@@ -42,14 +40,14 @@ def crawl_single_rss_feed(feed_id):
42
40
except RSSFeed .DoesNotExist :
43
41
raise Exception (f"RSS Feed with id { feed_id } not found" )
44
42
45
- logger .info (f"Starting to crawl RSS feed: { feed .name } ({ feed .url } )" )
43
+ logfire .info (f"Starting to crawl RSS feed: { feed .name } ({ feed .url } )" )
46
44
47
45
try :
48
46
# RSS 피드 파싱
49
47
parsed_feed = feedparser .parse (feed .url )
50
48
51
49
if parsed_feed .bozo :
52
- logger .warning (f"RSS feed { feed .name } has parsing issues: { parsed_feed .bozo_exception } " )
50
+ logfire .warning (f"RSS feed { feed .name } has parsing issues: { parsed_feed .bozo_exception } " )
53
51
54
52
new_items_count = 0
55
53
@@ -59,7 +57,7 @@ def crawl_single_rss_feed(feed_id):
59
57
link = getattr (entry , 'link' , '' )
60
58
61
59
if not guid and not link :
62
- logger .warning (f"Skipping entry without GUID or link in feed { feed .name } " )
60
+ logfire .warning (f"Skipping entry without GUID or link in feed { feed .name } " )
63
61
continue
64
62
65
63
# 중복 체크
@@ -98,10 +96,10 @@ def crawl_single_rss_feed(feed_id):
98
96
pub_date = pub_date
99
97
)
100
98
new_items_count += 1
101
- logger .debug (f"Created new RSS item: { rss_item .title } " )
99
+ logfire .debug (f"Created new RSS item: { rss_item .title } " )
102
100
103
101
except Exception as e :
104
- logger .error (f"Error creating RSS item for { link } : { str (e )} " )
102
+ logfire .error (f"Error creating RSS item for { link } : { str (e )} " )
105
103
continue
106
104
107
105
# 마지막 크롤링 시간 업데이트
@@ -114,7 +112,7 @@ def crawl_single_rss_feed(feed_id):
114
112
'total_entries' : len (parsed_feed .entries )
115
113
}
116
114
117
- logger .info (f"Completed crawling { feed .name } : { new_items_count } new items out of { len (parsed_feed .entries )} total entries" )
115
+ logfire .info (f"Completed crawling { feed .name } : { new_items_count } new items out of { len (parsed_feed .entries )} total entries" )
118
116
return result
119
117
120
118
except requests .RequestException as e :
0 commit comments