You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository was archived by the owner on Jul 16, 2024. It is now read-only.
Copy file name to clipboardExpand all lines: source/docs/getting-started/installation/sw_install/limelight.rst
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -20,5 +20,5 @@ Download the hardwareConfig.json file for the version of your Limelight:
20
20
21
21
:ref:`Import the hardwareConfig.json file <docs/hardware/config:Importing and Exporting Settings>`. Again, this is **REQUIRED** or target measurements will be incorrect, and LEDs will not work.
22
22
23
-
After installation you should be able to `locate the camera <https://web.archive.org/web/20220525051734/https://gloworm.vision//docs/quickstart/#finding-gloworm>`_ at: ``http://photonvision.local:5800/`` (not ``gloworm.local``, as previously)
23
+
After installation you should be able to `locate the camera <https://photonvision.github.io/gloworm-docs/docs/quickstart/#finding-gloworm>`_ at: ``http://photonvision.local:5800/`` (not ``gloworm.local``, as previously)
Copy file name to clipboardExpand all lines: source/docs/getting-started/installation/sw_install/romi.rst
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -16,6 +16,6 @@ Next, from the SSH terminal, run ``sudo nano /home/pi/runCamera`` then arrow dow
16
16
17
17
.. image:: images/nano.png
18
18
19
-
After it reboots, you should be able to `locate the PhotonVision UI <https://web.archive.org/web/20220525051734/https://gloworm.vision//docs/quickstart/#finding-gloworm>`_ at: ``http://10.0.0.2:5800/``.
19
+
After it reboots, you should be able to `locate the PhotonVision UI <https://photonvision.github.io/gloworm-docs/docs/quickstart/#finding-gloworm>`_ at: ``http://10.0.0.2:5800/``.
20
20
21
21
.. warning:: In order for settings, logs, etc. to be saved / take effect, ensure that PhotonVision is in writable mode.
Copy file name to clipboardExpand all lines: source/docs/getting-started/pipeline-tuning/about-pipelines.rst
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -24,7 +24,7 @@ AprilTag
24
24
25
25
This pipeline type is based on detecting AprilTag fiducial markers. More information about AprilTags can be found in the WPILib documentation. While being more performance intensive than the reflective and colored shape pipeline, it has the benefit of providing easy to use 3D pose information which allows localization.
26
26
27
-
.. note:: In order to get 3D Pose data about AprilTags, you are required to :ref:`calibrate your camera<docs/getting-started/pipeline-tuning/calibration:Calibration Steps>`.
27
+
.. note:: In order to get 3D Pose data about AprilTags, you are required to :ref:`calibrate your camera<docs/getting-started/pipeline-tuning/calibration:Calibrating Your Camera>`.
Copy file name to clipboardExpand all lines: source/docs/getting-started/pipeline-tuning/calibration.rst
+15-2Lines changed: 15 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -3,12 +3,13 @@ Calibrating Your Camera
3
3
4
4
.. important:: In order to detect AprilTags and use 3D mode, your camera must be calibrated at the desired resolution! Inaccurate calibration will lead to poor performance.
5
5
6
-
To calibrate a camera, images of a chessboard (or grid of dots) are taken. by comparing where the grid corners (or dots) should be in object space (for example, a dot once every inch in an 8x6 grid) with where they appear in the camera image, we can find a least-squares estimate for intrinsic camera properties like focal lengths, center point, and distortion coefficients. For more on camera calibration, please review the `OpenCV documentation <https://docs.opencv.org/4.x/dc/dbb/tutorial_py_calibration.html>`_.
6
+
To calibrate a camera, images of a chessboard (or grid of dots, or other target) are taken. by comparing where the grid corners (or dots) should be in object space (for example, a dot once every inch in an 8x6 grid) with where they appear in the camera image, we can find a least-squares estimate for intrinsic camera properties like focal lengths, center point, and distortion coefficients. For more on camera calibration, please review the `OpenCV documentation <https://docs.opencv.org/4.x/dc/dbb/tutorial_py_calibration.html>`_.
7
7
8
8
.. warning:: While any resolution can be calibrated, resolutions lower than 960x720 are often too low to provide accurate results. Additionally, high resolutions may be too performance intensive for a coprocessor like a Raspberry Pi to handle (solutions to this are being looked into). Thus, we recommend 960x720 when using 3D mode.
9
9
10
10
.. note::The calibration data collected during calibration is specific to each physical camera, as well as each individual resolution.
11
11
12
+
12
13
Calibration Tips
13
14
----------------
14
15
Accurate camera calibration is required in order to get accurate pose measurements when using AprilTags and 3D mode. The tips below should help ensure success:
@@ -36,7 +37,19 @@ Accurate camera calibration is required in order to get accurate pose measuremen
36
37
Following the ideas above should help in getting an accurate calibration.
37
38
38
39
Calibration Steps
39
-
-----------------
40
+
=================
41
+
42
+
Your camera can be calibrated using either the utility built into PhotonVision, which performs all the calculations on your coprocessor, or using a website such as `calibdb <https://calibdb.net/>`, which uses a USB webcam connected to your laptop. The integrated calibration utility is currently the only one that works with ribbon-cable CSI cameras or Limelights, but for USB webcams, calibdb is the preferred option.
43
+
44
+
Calibrating using calibdb
45
+
-------------------------
46
+
47
+
Calibdb uses a modified chessboard/aruco marker combination target called `ChArUco targets <https://docs.opencv.org/3.4/df/d4a/tutorial_charuco_detection.html>`. The website currently only supports Chrome browser.
48
+
49
+
Download and print out (or display on a monitor) the calibration by clicking Show Pattern. Click "Calibrate" and align your camera with the ghost overlay of the calibration board. The website automatically calculates the next position and displays it for you. When complete, download the calibration (do **not** use the OpenCV format). Reconnect your camera to your coprocessor and navigate to the PhotonVision web interface's camera tab. Ensure the correct camera is selected, and click the "Import from CalibDB" button. Your calibration data will be automatically saved and applied!
Copy file name to clipboardExpand all lines: source/docs/getting-started/pipeline-tuning/reflectiveAndShape/3D.rst
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,7 +1,7 @@
1
1
3D Tuning
2
2
=========
3
3
4
-
In 3D mode, the SolvePNP algorithm is used to compute the position and rotation of the target relative to the robot. This requires your :ref:`camera to be calibrated <docs/getting-started/pipeline-tuning/calibration:Calibration Steps>` which can be done through the cameras tab.
4
+
In 3D mode, the SolvePNP algorithm is used to compute the position and rotation of the target relative to the robot. This requires your :ref:`camera to be calibrated <docs/getting-started/pipeline-tuning/calibration:Calibrating Your Camera>` which can be done through the cameras tab.
5
5
6
6
The target model dropdown is used to select the target model used to compute target position. This should match the target your camera will be tracking.
Copy file name to clipboardExpand all lines: source/index.rst
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,7 +1,7 @@
1
1
.. image:: assets/PhotonVision-Header-onWhite.png
2
2
:alt:PhotonVision
3
3
4
-
Welcome to the official documentation of PhotonVision! PhotonVision is the free, fast, and easy-to-use vision processing solution for the *FIRST*\ Robotics Competition. PhotonVision is designed to get vision working on your robot *quickly*, without the significant cost of other similar solutions. PhotonVision supports a variety of COTS hardware, including the Raspberry Pi 3 and 4, the `Gloworm smart camera <https://web.archive.org/web/20220525051734/https://gloworm.vision//>`_, and the `SnakeEyes Pi hat <https://www.playingwithfusion.com/productview.php?pdid=133>`_.
4
+
Welcome to the official documentation of PhotonVision! PhotonVision is the free, fast, and easy-to-use vision processing solution for the *FIRST*\ Robotics Competition. PhotonVision is designed to get vision working on your robot *quickly*, without the significant cost of other similar solutions. PhotonVision supports a variety of COTS hardware, including the Raspberry Pi 3 and 4, the `Gloworm smart camera <https://photonvision.github.io/gloworm-docs/docs/quickstart/#finding-gloworm>`_, and the `SnakeEyes Pi hat <https://www.playingwithfusion.com/productview.php?pdid=133>`_.
0 commit comments