Skip to content

Commit 2ab0a3b

Browse files
committed
fix celery worker log name
1 parent f310d7e commit 2ab0a3b

File tree

3 files changed

+3
-19
lines changed

3 files changed

+3
-19
lines changed

deploy_prod.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -64,4 +64,4 @@ gunicorn --workers=2 \
6464
celery -A pythonkr_backend multi start worker1 -c2 -B \
6565
--loglevel=info \
6666
--pidfile=/home/pk/celery-%n.pid \
67-
--logfile=/home/pk/logs/celery-%n%I.log
67+
--logfile=/home/pk/logs/celery-%n-%i.log

pythonkr_backend/curation/tasks.py

Lines changed: 0 additions & 18 deletions
Original file line numberDiff line numberDiff line change
@@ -7,9 +7,6 @@
77
from django.utils import timezone as django_timezone
88
from django.core.files.base import ContentFile
99
from .models import RSSFeed, RSSItem
10-
import logging
11-
12-
logger = logging.getLogger(__name__)
1310

1411
def crawl_all_rss_feeds():
1512
"""모든 활성화된 RSS 피드를 크롤링합니다."""
@@ -28,12 +25,10 @@ def crawl_all_rss_feeds():
2825
results['processed_feeds'] += 1
2926
results['new_items'] += result.get('new_items', 0)
3027
logfire.info(f"Successfully crawled feed {feed.name}: {result.get('new_items', 0)} new items")
31-
logger.info(f"Successfully crawled feed {feed.name}: {result.get('new_items', 0)} new items")
3228
except Exception as e:
3329
error_msg = f"Error crawling feed {feed.name}: {str(e)}"
3430
results['errors'].append(error_msg)
3531
logfire.error(error_msg)
36-
logger.error(error_msg)
3732

3833
return results
3934

@@ -46,15 +41,13 @@ def crawl_single_rss_feed(feed_id):
4641
raise Exception(f"RSS Feed with id {feed_id} not found")
4742

4843
logfire.info(f"Starting to crawl RSS feed: {feed.name} ({feed.url})")
49-
logger.info(f"Starting to crawl RSS feed: {feed.name} ({feed.url})")
5044

5145
try:
5246
# RSS 피드 파싱
5347
parsed_feed = feedparser.parse(feed.url)
5448

5549
if parsed_feed.bozo:
5650
logfire.warning(f"RSS feed {feed.name} has parsing issues: {parsed_feed.bozo_exception}")
57-
logger.warning(f"RSS feed {feed.name} has parsing issues: {parsed_feed.bozo_exception}")
5851

5952
new_items_count = 0
6053

@@ -65,7 +58,6 @@ def crawl_single_rss_feed(feed_id):
6558

6659
if not guid and not link:
6760
logfire.warning(f"Skipping entry without GUID or link in feed {feed.name}")
68-
logger.warning(f"Skipping entry without GUID or link in feed {feed.name}")
6961
continue
7062

7163
# 중복 체크
@@ -105,11 +97,9 @@ def crawl_single_rss_feed(feed_id):
10597
)
10698
new_items_count += 1
10799
logfire.debug(f"Created new RSS item: {rss_item.title}")
108-
logger.debug(f"Created new RSS item: {rss_item.title}")
109100

110101
except Exception as e:
111102
logfire.error(f"Error creating RSS item for {link}: {str(e)}")
112-
logger.error(f"Error creating RSS item for {link}: {str(e)}")
113103
continue
114104

115105
# 마지막 크롤링 시간 업데이트
@@ -123,7 +113,6 @@ def crawl_single_rss_feed(feed_id):
123113
}
124114

125115
logfire.info(f"Completed crawling {feed.name}: {new_items_count} new items out of {len(parsed_feed.entries)} total entries")
126-
logger.info(f"Completed crawling {feed.name}: {new_items_count} new items out of {len(parsed_feed.entries)} total entries")
127116
return result
128117

129118
except requests.RequestException as e:
@@ -136,15 +125,13 @@ def crawl_single_rss_feed(feed_id):
136125
def crawl_rss():
137126
"""10분마다 실행되는 RSS 크롤링 태스크"""
138127
logfire.info("start to crawl rss")
139-
logger.info("start to crawl rss")
140128
return crawl_all_rss_feeds()
141129

142130

143131
@shared_task
144132
def crawl_rss_item_content():
145133
"""RSS 아이템의 본문을 크롤링하는 태스크 (10분마다 실행)"""
146134
logfire.info("Starting RSS item content crawling")
147-
logger.info("Starting RSS item content crawling")
148135

149136
# 2주 이내의 크롤링되지 않은 최신 1개 아이템 가져오기
150137
two_weeks_ago = django_timezone.now() - timedelta(days=14)
@@ -155,11 +142,9 @@ def crawl_rss_item_content():
155142

156143
if not pending_item:
157144
logfire.info("No pending RSS items to crawl")
158-
logger.info("No pending RSS items to crawl")
159145
return {"status": "no_items", "message": "No pending items to crawl"}
160146

161147
logfire.info(f"Crawling RSS item: {pending_item.title} ({pending_item.link})")
162-
logger.info(f"Crawling RSS item: {pending_item.title} ({pending_item.link})")
163148

164149
# 크롤링 상태를 진행 중으로 변경 (동시 처리 방지)
165150
pending_item.crawling_status = 'completed' # 임시로 설정하여 중복 처리 방지
@@ -186,7 +171,6 @@ def crawl_rss_item_content():
186171
pending_item.save()
187172

188173
logfire.info(f"Successfully crawled RSS item: {pending_item.title}")
189-
logger.info(f"Successfully crawled RSS item: {pending_item.title}")
190174

191175
return {
192176
"status": "success",
@@ -202,7 +186,6 @@ def crawl_rss_item_content():
202186
pending_item.save(update_fields=['crawling_status', 'error_message'])
203187

204188
logfire.error(error_msg)
205-
logger.error(error_msg)
206189

207190
return {
208191
"status": "failed",
@@ -217,7 +200,6 @@ def crawl_rss_item_content():
217200
pending_item.save(update_fields=['crawling_status', 'error_message'])
218201

219202
logfire.error(error_msg)
220-
logger.error(error_msg)
221203

222204
return {
223205
"status": "failed",

pythonkr_backend/pythonkr_backend/celery.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -24,8 +24,10 @@
2424
def init_worker(*args, **kwargs):
2525
logfire.configure(service_name="celery-worker", send_to_logfire='if-token-present')
2626
logfire.instrument_celery()
27+
logfire.instrument_pydantic_ai()
2728

2829
@beat_init.connect()
2930
def init_beat(*args, **kwargs):
3031
logfire.configure(service_name="celery-beat", send_to_logfire='if-token-present')
3132
logfire.instrument_celery()
33+
logfire.instrument_pydantic_ai()

0 commit comments

Comments
 (0)