-
Notifications
You must be signed in to change notification settings - Fork 2
CSR report
This report outlines the CSR impact of our A.P.RI.L. project. The project enables users to efficiently gather and analyze web data through a user-friendly interface. We have ensured the project adheres to ethical standards of data collection, minimizes environmental impact, and provides social value through transparency and accessibility.
- Executive Summary
- Introduction to the IT Project
- CSR Strategy and Goals
- ESG Considerations
- Key Metrics and Performance Indicators
- Challenges and Solutions
- Future CSR Initiatives
- Conclusion
Analysis of Coastal Risk Perception in Occitanie (A.P.Ri.L: Analyse de la Perception des Risques Littoraux en Occitanie)
Coastal areas are facing increasing risks due to climate change, including rising sea levels, coastal erosion, and extreme weather events such as storms and flooding. Understanding how various stakeholders—such as local communities, policy makers, businesses, and environmental organizations—perceive these risks is crucial for informed decision-making and effective coastal management strategies. The A.P.Ri.L. (Analyse de la Perception des Risques Littoraux en Occitanie) project seeks to fill this gap by conducting a detailed analysis of coastal risk perception in the Occitanie region, located in southern France, which has a significant and vulnerable coastline.
The APRIL project aims to collect a representative corpus of textual data from diverse web-based sources, including news articles, government reports, blog posts, social media, and scientific literature. This data collection is intended to capture a wide range of perspectives on coastal risks—such as erosion, flooding, and habitat loss—within the region. By gathering this data, the project aspires to develop a comprehensive understanding of the concerns, priorities, and awareness levels of different stakeholder groups.
Using advanced Natural Language Processing (NLP) techniques, the project will analyze the collected data to identify key themes, concerns, and the emotional tone of discourse around coastal risks. For instance, it will evaluate whether stakeholders view coastal risks as urgent or manageable, what measures they believe are necessary, and how they perceive the role of government or private actors in managing these risks. By structuring and analyzing this information, the project will provide valuable insights into the perceptions of coastal risks at both local and regional levels.
The goal of the APRIL project is not just to create a static snapshot of risk perception but to offer an evolving analysis that can inform adaptive coastal management policies. Insights derived from the project will help guide local authorities in designing and implementing policies that reflect the needs and concerns of the population. Furthermore, this project contributes to greater public awareness of coastal risks by highlighting areas where perception might differ from scientific risk assessments, thus helping to bridge the gap between science, policy, and public understanding.
As part of its methodology, the APRIL project emphasizes the ethical use of data, ensuring that the sources of information are transparent and respect data privacy regulations. Data collected from public sources will be handled with strict adherence to ethical guidelines to avoid infringing on the privacy of individuals or violating the terms of service of websites.
Ultimately, by combining web scraping, data analysis, and stakeholder engagement, the APRIL project provides a robust framework for understanding coastal risk perception in Occitanie. The findings will be crucial for informing both short-term interventions, such as emergency preparedness, and long-term strategies, such as coastal zone management and land-use planning, aimed at reducing vulnerability to coastal risks.
This report outlines the CSR impact of our web scraping pipeline project. The project enables users to efficiently gather and analyze web data through a user-friendly interface. We have ensured the project adheres to ethical standards of data collection, minimizes environmental impact, and provides social value through transparency and accessibility.
This IT project involves collecting web-based data through a web scraping pipeline. The pipeline automates the collection of reviews and information on specific topics, allowing for the efficient gathering of unstructured data from various websites. This data is subsequently processed to create a structured corpus, which is then analyzed using Natural Language Processing (NLP) techniques.
The entire pipeline is local, meaning no data is shared outside the project environment. The goal is to use this data for academic or research purposes, allowing for insights into the chosen topics.
A user-friendly interface (UI) is developed to interact with the scraped data, allowing researchers and analysts to easily visualize, query, and analyze the corpus. The UI ensures a seamless experience for end-users, promoting efficient access and manipulation of the data collected.
-
Efficient data collection and processing. The primary objective is to develop a reliable system for scraping large amounts of data from diverse web sources while adhering to ethical guidelines. The data is processed to build a structured, clean corpus for analysis, facilitating valuable insights through NLP.
-
User-friendly interaction with the data through a UI. Another key objective is to provide users (e.g., researchers, data analysts) with a simple and effective way to explore and manipulate the data collected. The UI enhances the accessibility and usability of the data, even for those with limited technical expertise.
-
Compliance with Data Privacy Regulations: The project ensures compliance with global and regional data privacy regulations, such as the General Data Protection Regulation (GDPR), Web scraping operations have to respect the terms and conditions of the websites being scraped, ensuring that only publicly available data is collected. No personal or sensitive data is scraped, stored, or processed, thereby protecting individual privacy and complying with relevant legal frameworks.
Evidence of Compliance:
- Scraping only public, non-personal information.
- Regular auditing of data collection processes to ensure compliance with laws and website policies.
- Implementing safeguards like data anonymization and strict data access controls.
-
Transparent Data Collection: Transparency is a fundamental aspect of the project. All data collection activities are clearly documented, ensuring that stakeholders are aware of the sources and methods used to collect the data. The project team acknowledges the ownership of data by its original sources and refrains from using any proprietary data without permission.
Best Practices:
- Maintaining detailed logs of websites and data scraped.
- Clear documentation of the data collection process for transparency and accountability.
-
Consent and Permissions: While web scraping often involves gathering publicly available information, it is important to respect each website's robots.txt file or scraping policies. If a website explicitly prohibits scraping, the project does not collect data from it. For data sources where terms are ambiguous, the project seeks permission to scrape.
Examples:
- Scraping only from sites that permit data extraction in their terms of service.
- Seeking explicit consent when necessary.
-
Efficient Resource Usage: The project is designed to minimize its environmental footprint by optimizing resource usage. The web scraping pipeline is designed to be computationally efficient, reducing the energy consumption associated with data collection and processing.
Steps to Optimize Efficiency:
- Using low-power servers or cloud computing resources with low carbon footprints.
- Efficient data processing techniques to reduce unnecessary computation.
- Minimizing bandwidth consumption during data scraping to reduce network congestion.
-
Green IT Practices: Green IT practices are embedded into the project’s operations. Wherever possible, energy-efficient hardware and cloud services that are powered by renewable energy sources are utilized. The local storage infrastructure is also optimized for energy savings.
Examples of Green Practices:
- Using data centers that utilize renewable energy sources.
- Reducing storage redundancy and avoiding overprovisioning in servers.
- Regularly auditing the environmental impact of hardware used for scraping and processing.
-
Access to Data for Social Good: The data collected and analyzed through this project can potentially be used for social good. By providing insights into specific topics, it could aid in research aimed at solving pressing societal issues. This data can help in public health research, consumer behavior analysis, and policy-making, depending on the topic of focus.
Impact Example:
- Use of the corpus for research into topics such as misinformation, healthcare trends, or social justice issues.
- Potential sharing of anonymized insights or aggregate-level data with non-profits or public institutions for better decision-making.
-
Job Creation and Skills Development: The project offers opportunities for job creation, particularly in the fields of data science, machine learning, and software development. By building the web scraping pipeline and the associated NLP models, it provides team members with the opportunity to upskill in areas such as data collection, processing, and analysis. It also supports the creation of new jobs in IT, analytics, and environmental science.
Contributions:
- Internships or apprenticeships for students and early-career professionals.
- Workshops or courses on web scraping and NLP offered to communities and academic institutions.
-
Accessibility: The project aims to ensure that the data and insights derived from the NLP analysis are accessible to a broad range of users. The user interface is designed with accessibility in mind, ensuring it is intuitive and usable for people of diverse technical backgrounds, including those with disabilities.
Accessibility Features:
- The UI includes features like screen reader support, keyboard navigation, and customizable settings (e.g., text size, contrast).
- Documentation and tutorials to assist non-technical users in utilizing the platform.
- Energy Efficiency:
- Cloud Provider's Environmental Impact:
- Ethical Web Scraping: a. Compliance with Legal Standards:
Data Privacy Considerations: Ensure that your scraping pipeline respects privacy laws such as the General Data Protection Regulation (GDPR) in Europe. When scraping personal data, ensuring that no personally identifiable information (PII) is processed or stored without proper consent is crucial.
b. Minimizing Negative Impact on Websites:
Limiting server load: Ethical scraping also involves limiting the frequency and volume of requests to avoid overloading servers. Respect rate limits to prevent disrupting the normal functioning of the websites you're scraping.
Data Integrity: Scraping responsibly also means not using the data for any form of malicious intent or misinformation. The data should be used in ways that align with your institution's research goals, contributing to legitimate, constructive research.
c. Transparency and Collaboration:
Credit to Data Sources: Whenever possible, credit the websites or data sources from which you collected information. This fosters a culture of respect and transparency.
- Stakeholder Impact:
a. Benefiting End Users and the Public:
Advancing Scientific Knowledge for Ecological Evolution: Although the raw data collected is private and restricted to your research project, the insights derived from your analysis will be shared as part of a larger public study, contributing to ecological and environmental research on a global scale. This ensures that while the integrity and confidentiality of the data are maintained, the research outcomes benefit society by enhancing understanding of ecological trends and supporting environmental sustainability.
Informing Policy and Global Action: Your research conclusions can be used in the development of national and international environmental policies. By contributing to a larger public study, your analysis helps influence governmental and organizational decisions on matters like climate change, habitat preservation, and sustainable practices, which will have widespread positive effects on both ecosystems and human populations.
Supporting Public Awareness and Educational Initiatives: Once your conclusions become part of the broader public study, they can be used in educational materials or media reports to raise awareness about critical ecological issues. This could lead to more informed public discourse and potentially inspire grassroots environmental activism or support for green policies.
b. Impact on Website Owners:
Collaboration with Data Providers: Although the raw data remains private, the larger public study to which your research contributes will likely collaborate with ecological data sources and organizations. By responsibly handling private data and contributing only the conclusions, you ensure that your research remains aligned with ethical standards while supporting the greater mission of environmental sustainability.
- Data Governance Policies:
Home | Contributors | Report an Issue | Licence
© 2024 APRIL. | Version 1.0 | Last updated on 2025-01-14