Skip to content

Conversation

@nancyboukamel-ds
Copy link

No description provided.

@addie9800 addie9800 self-assigned this Oct 26, 2025
Copy link
Collaborator

@addie9800 addie9800 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks for adding the first Lebanese publisher. Unfortunately, the code you provided does not run. I have commented on the relevant section, so that you can go ahead and fix it.

domain = "https://www.lbcgroup.tv",
parser = LBCGroupParser,
sources=[
RSSFeed("https://www.lbcgroup.tv/Rss/latest-news/en"),
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Since the default language is arabic, please add the following parameter to to each English source languages={"en"}. Also, since there is an arabic RSSFeed, please add that as well: RSSFeed("https://www.lbcgroup.tv/Rss/latest-news/ar")

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I think we can add all RSSFeeds from https://www.lbcgroup.tv/rss/ar and https://www.lbcgroup.tv/rss/en

parser = LBCGroupParser,
sources=[
RSSFeed("https://www.lbcgroup.tv/Rss/latest-news/en"),
NewsMap("https://www.lbcgroup.tv/newssitemap.xml"),
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This this source is multi-lingual, please override the language with languages={"en", "ar"}

sources=[
RSSFeed("https://www.lbcgroup.tv/Rss/latest-news/en"),
NewsMap("https://www.lbcgroup.tv/newssitemap.xml"),
Sitemap("https://www.lbcgroup.tv/sitemap.xml"),
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I think this sitemap can safely be removed, at least I didn't find any articles that weren't also part of the Newsmap. And more problematically, there were numerous pages that were not relevant to Fundus.

# Use the defined content_selector to locate the block of text.
return extract_article_body_with_selector(
self.precomputed.doc,
content_selector=self._content_container_selector,
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This line does not really make sense and causes the program to crash. You have defined the selector as content_container_selector and try to access it as _content_container_selector. Also, extract_article_body_with_selector has no parameter content_selector.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants