GSoC 2025 Proposal 4: DeepForest Vision Agent connection with LandingAI #973
Replies: 4 comments 17 replies
-
|
Hi @Samia35-2973, Thank you for reaching out and for your interest in contributing to DeepForest through GSoC 2025!
Base on the project details as above, Here is an example of the demo from the landing-ai service given a text query of
Try out all DeepForest models and their detection classes to understand how they could be leveraged for a user-friendly interface that selects new images based on agent responses. Think about how we can design a user-friendly interface that allows users to input text-based queries, load a model, and detect objects in images based on the query. |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
|
Hi. I've added a basic integration of the Gemini model with the DeepForest tool in this pull request: weecology/deepforest-agent#2. The goal is to test whether Gemini can effectively interact with and call DeepForest tools. I've also included the current workflow diagram in the README file. Could you please run and test it out? |
Beta Was this translation helpful? Give feedback.




Uh oh!
There was an error while loading. Please reload this page.
-
Hi Henry, Ben, and Ethan,
I am Samia Haque, a recent graduate with a B.Sc. in Software Engineering (major in Data Science). I have been actively working on integrating computer vision with post-hoc interpretability techniques and LLM-based automation, exploring how these technologies can enhance decision-making in various domains.
My Skills and Experience
Object Detection & Post-hoc AI Explainability:
I am currently working on multi-stage traffic anomaly detection, where I utilized YOLOv9 and YOLOv10 to detect traffic congested regions from images. I am also working on its post-hoc interpretability and LLM integration.
LLM Integration & AI Automation:
I have experience in prompt engineering, dataset curation, and human feedback-driven improvements for LLMs. While I am not deeply involved in LLM fine-tuning or RLHF at a model-training level, my work in structuring datasets and evaluating LLM outputs aligns with Vision Agent’s AI-driven image analysis pipeline.
I am excited to contribute to GSoC 2025, particularly through Proposal 4: DeepForest Vision Agent connection with LandingAI because it combines object detection with AI-driven automation. The idea of integrating a Vision Agent for automated labeling and active learning complements my ongoing exploration of AI-driven decision-making systems. Currently, I am studying the DeepForest codebase, documentation, and related research papers to better understand its architecture.
I would appreciate guidance on recommended resources or discussions for better understanding Vision Agent’s integration.
Thanks,
Samia
Contact Details:
Beta Was this translation helpful? Give feedback.
All reactions