Skip to content

Commit 321b825

Browse files
authored
Merge pull request #100 from FocoosAI/98-docs-fix-missing-links-in-documentation
Refactor model documentation and navigation structure
2 parents 3a5c922 + 74e9aec commit 321b825

22 files changed

+496
-5918
lines changed

README.md

Lines changed: 8 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -1,15 +1,17 @@
11
![Tests](https://github.com/FocoosAI/focoos/actions/workflows/test.yml/badge.svg??event=push&branch=main)
22
[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/FocoosAI/focoos/blob/main/tutorials/training.ipynb)
3+
[![Documentation](https://img.shields.io/badge/docs-latest-blue)](https://focoosai.github.io/focoos/)
4+
35
# Welcome to Focoos AI
46

57
Focoos AI provides an advanced development platform designed to empower developers and businesses with efficient, customizable computer vision solutions. Whether you're working with data from cloud infrastructures or deploying on edge devices, Focoos AI enables you to select, fine-tune, and deploy state-of-the-art models optimized for your unique needs.
68

7-
## SDK Overview
9+
## Overview
810

9-
<!-- Unlock the full potential of Focoos AI with the Focoos Python SDK! 🚀 -->
10-
The Focoos Python SDK is your gateway to easily access cutting-edge computer vision models and development tools. With just a few lines of code, you can **fine tune** pre-trained models tailored to your specific needs.
11+
<!-- Unlock the full potential of Focoos AI with the Focoos library! 🚀 -->
12+
The Focoos library is your gateway to easily access cutting-edge computer vision models and development tools. With just a few lines of code, you can **fine tune** pre-trained models tailored to your specific needs.
1113

12-
Whether you're working in the cloud or on edge devices, the Focoos Python SDK seamlessly integrates into your workflow, accelerating development and simplifying the implementation of computer vision solutions.
14+
Whether you're working in the cloud or on edge devices, the Focoos library seamlessly integrates into your workflow, accelerating development and simplifying the implementation of computer vision solutions.
1315

1416
### Key Features 🔑
1517

@@ -27,7 +29,7 @@ Whether you're working in the cloud or on edge devices, the Focoos Python SDK se
2729
Ready to dive in? Get started with the setup in just a few simple steps!
2830

2931
## Installation
30-
**Install** the Focoos Python SDK (for more options, see [setup](https://focoosai.github.io/focoos/setup))
32+
**Install** the Focoos library (for more options, see [setup](https://focoosai.github.io/focoos/setup))
3133

3234
```bash linenums="0"
3335
uv pip install 'focoos @ git+https://github.com/FocoosAI/focoos.git'
@@ -105,7 +107,7 @@ Using Focoos AI helps you save both time and money while delivering high-perform
105107
- **4x Cheaper** 💰: Our models require up to 4x less computational power, letting you save on hardware or cloud bill while ensuring high-quality results.
106108
- **Tons of CO2 saved annually per model** 🌱: Our models are energy-efficient, helping you reduce your carbon footprint by using less powerful hardware with respect to mainstream models.
107109

108-
See the list of our models in the [models](https://focoosai.github.io/focoos/models/) section.
110+
See the list of our models in the [models](https://focoosai.github.io/focoos/models/models) section.
109111

110112
---
111113
### Start now!

docs/concepts.md

Lines changed: 9 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@ The `FocoosModel` class is the main interface for working with computer vision m
1515

1616
### Loading Strategies
1717

18-
The primary method for loading models is using the `ModelManager.get()` (see [`ModelManager`](../api/model_manager/#focoos.model_manager.ModelManager)). It supports multiple loading strategies based on the input parameters. The return value is a [Focoos Model](#focoosmodel).
18+
The primary method for loading models is using the `ModelManager.get()` (see [`ModelManager`](/focoos/api/model_manager/#focoos.model_manager.ModelManager)). It supports multiple loading strategies based on the input parameters. The return value is a [Focoos Model](#focoosmodel).
1919

2020
The ModelManager employs different loading strategies based on the input:
2121

@@ -102,7 +102,7 @@ model = ModelManager.get(
102102

103103
#### 4. From ModelInfo Object
104104

105-
The [`ModelInfo`](../api/ports/#focoos.ports.ModelInfo) class represents comprehensive model metadata including architecture specifications, training configuration, class definitions, and performance metrics. This method provides the most programmatic control over model instantiation.
105+
The [`ModelInfo`](/focoos/api/ports/#focoos.ports.ModelInfo) class represents comprehensive model metadata including architecture specifications, training configuration, class definitions, and performance metrics. This method provides the most programmatic control over model instantiation.
106106

107107
**When to use**: Programmatically construct models, work with dynamic configurations, integrate with custom model management systems, or when you need fine-grained control over model instantiation.
108108

@@ -142,7 +142,7 @@ Performs end-to-end inference on input images with automatic preprocessing and p
142142
- NumPy arrays (`numpy.ndarray`)
143143
- PyTorch tensors (`torch.Tensor`)
144144

145-
The input images are automatically preprocessed to the correct size and format required by the model. After inference, the raw model outputs are postprocessed into a standardized [`FocoosDetections`](../api/ports/#focoos.ports.FocoosDetections) format that provides easy access to:
145+
The input images are automatically preprocessed to the correct size and format required by the model. After inference, the raw model outputs are postprocessed into a standardized [`FocoosDetections`](/focoos/api/ports/#focoos.ports.FocoosDetections) format that provides easy access to:
146146

147147
- Detected object classes and confidence scores
148148
- Bounding box coordinates
@@ -155,7 +155,7 @@ This provides a simple, unified interface for running inference regardless of th
155155
- `inputs`: Input images in various supported formats (`PIL.Image.Image`, `numpy.ndarray`, `torch.Tensor`)
156156
- `**kwargs`: Additional arguments passed to postprocessing
157157

158-
**Returns:** [`FocoosDetections`](../api/ports/#focoos.ports.FocoosDetections) containing detection/segmentation results
158+
**Returns:** [`FocoosDetections`](/focoos/api/ports/#focoos.ports.FocoosDetections) containing detection/segmentation results
159159

160160
**Example:**
161161
```python
@@ -176,7 +176,7 @@ for detection in detections.detections:
176176
### Training
177177
Trains the model on provided datasets. The training function accepts:
178178

179-
- `args`: Training configuration ([TrainerArgs](../api/ports/#focoos.ports.TrainerArgs)) specifying the main hyperparameters, among which:
179+
- `args`: Training configuration ([TrainerArgs](/focoos/api/ports/#focoos.ports.TrainerArgs)) specifying the main hyperparameters, among which:
180180
- `run_name`: Name for the training run
181181
- `output_dir`: Name for the output folder
182182
- `num_gpus`: Number of GPUs to use (must be >= 1)
@@ -186,7 +186,7 @@ Trains the model on provided datasets. The training function accepts:
186186
- `data_val`: Validation dataset (MapDataset)
187187
- `hub`: Optional FocoosHUB instance for experiment tracking
188188

189-
The data can be obtained using the [AutoDataset](../api/auto_dataset/#focoos.data.auto_dataset.AutoDataset) helper.
189+
The data can be obtained using the [AutoDataset](/focoos/api/auto_dataset/#focoos.data.auto_dataset.AutoDataset) helper.
190190

191191
After the training is complete, the model will have updated weights and can be used for inference or export. Furthermore, in the `output_dir` can be found the model metadata (`model_info.json`) and the PyTorch weights (`model_final.pth`).
192192

@@ -214,7 +214,7 @@ model.train(train_args, train_dataset, val_dataset, hub=hub)
214214
### Model Export
215215

216216
Exports the model to different runtime formats for optimized inference. The main function arguments are:
217-
- `runtime_type`: specify the target runtime and must be one of the supported (see [RuntimeType](../api/ports/#focoos.ports.RuntimeType))
217+
- `runtime_type`: specify the target runtime and must be one of the supported (see [RuntimeType](/focoos/api/ports/#focoos.ports.RuntimeType))
218218
- `out_dir`: the destination folder for the exported model
219219
- `image_size`: the target image size, as an optional integer
220220

@@ -246,7 +246,7 @@ The `InferModel` class represents an optimized model for inference, typically cr
246246

247247
### Initialization
248248

249-
InferModel instances are typically created through the `export()` method of a [FocoosModel](#focoosmodel), which handles the model optimization and conversion process. This method allows you to specify the target runtime (see the availables in [`Runtimetypes`](focoos/api/ports/#focoos.ports.RuntimeType)) and the output directory for the exported model. The `export()` method returns an `InferModel` instance that is optimized for fast and efficient inference.
249+
InferModel instances are typically created through the `export()` method of a [FocoosModel](#focoosmodel), which handles the model optimization and conversion process. This method allows you to specify the target runtime (see the availables in [`Runtimetypes`](/focoos/api/ports/#focoos.ports.RuntimeType)) and the output directory for the exported model. The `export()` method returns an `InferModel` instance that is optimized for fast and efficient inference.
250250

251251
**Example:**
252252
```python
@@ -281,7 +281,7 @@ This provides a simple, unified interface for running inference regardless of th
281281
- `inputs`: Input images in various supported formats (`PIL.Image.Image`, `numpy.ndarray`, `torch.Tensor`)
282282
- `**kwargs`: Additional arguments passed to postprocessing
283283

284-
**Returns:** [`FocoosDetections`](../api/ports/#focoos.ports.FocoosDetections) containing detection/segmentation results
284+
**Returns:** [`FocoosDetections`](/focoos/api/ports/#focoos.ports.FocoosDetections) containing detection/segmentation results
285285

286286
**Example:**
287287
```python

docs/hub/overview.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,8 @@
11
# 🚀 Focoos HUB Overview
22
The Focoos HUB is a cloud-based platform that provides seamless integration between your local development environment and the Focoos AI ecosystem. It enables you to manage models, datasets, perform remote inference operations, and monitor training progress through a unified API.
33

4+
[![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/FocoosAI/focoos/blob/main/tutorials/hub.ipynb)
5+
46
## What is Focoos HUB?
57

68
Focoos HUB serves as your gateway to:

docs/inference.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -48,7 +48,7 @@ In this section, you'll run a model on the Focoos' servers instead of on your ma
4848

4949
```python
5050
model_ref = "<YOUR-MODEL-REF>"
51-
dataset = hub.get_remote_model(model_ref)
51+
model = hub.get_remote_model(model_ref)
5252
```
5353

5454
Using the model is as simple as it could! Just call it with an image.
@@ -139,7 +139,7 @@ In the following cells, we will export the previous model for one of these and r
139139

140140
### Torchscript
141141

142-
We already provide multiple inference runtime, that you can see on the [`RuntimeTypes`](focoos/api/ports/#focoos.ports.RuntimeType) enum. Let's select Torchscript as an example.
142+
We already provide multiple inference runtime, that you can see on the [`RuntimeTypes`](/focoos/api/ports/#focoos.ports.RuntimeType) enum. Let's select Torchscript as an example.
143143

144144
```python
145145
from focoos.ports import RuntimeType
@@ -169,7 +169,7 @@ But, let's see its latency, that should be substantially lower than the pure pyt
169169
optimized_model.benchmark(iterations=10, size=512)
170170
```
171171

172-
You can use different runtimes that may fit better your device, such as TensorRT. See the list of available Runtimes at [`RuntimeTypes`](focoos/api/ports/#focoos.ports.RuntimeType). Please note that you need to install the relative packages for onnx and tensorRT for using them.
172+
You can use different runtimes that may fit better your device, such as TensorRT. See the list of available Runtimes at [`RuntimeTypes`](/focoos/api/ports/#focoos.ports.RuntimeType). Please note that you need to install the relative packages for onnx and tensorRT for using them.
173173

174174
### ONNX with TensorRT
175175
```python

0 commit comments

Comments
 (0)