6328 (TBA | Github | Website)'s Advanced Swerve Drive Project with More Enhanced Physics Simulation (Collision Detection and More)
π We extend our deepest appreciation to Team 6328 for their open source project, which have made this project possible.
βοΈ This project is based on an example from AdvantageKit, which is under the GPLv3 license. All teams are welcome to use it, provided that you adhere to the Advantage Kit License.
The standout feature of this project is the integration of the dyn4j physics engine, which allows for the creation of a highly realistic and interactive simulation environment.
In our simulator, the robot can collide with field obstacles, interact with opposing robots, and engage with game pieces, providing a dynamic and immersive experience.
Previous swerve drive simulations modeled the drivetrain using several DCMotorSim instances. In contrast, this project's drivetrain simulation calculates the frictional and propelling forces on each swerve module. These forces are then used to model the drivetrain as a rigid body within the physics engine. [Source Code]
In our simulation, game pieces on the field have collision boundaries and can interact with the robot. The simulator also supports a fixed intake module on the robot, enabling it to automatically grab game pieces upon contact. [Source Code | Example Code]
Simulated opponent robots can be either manually controlled using a gamepad for defensive play or set to follow pre-programmed cycle paths. Like real robots, these opponents have collision boundaries, allowing drivers to practice both defensive and offensive strategies effectively. [Source Code]
By modeling the interaction between the wheels and the floor, our simulation can replicate odometry measurement errors caused by skidding.
Additionally, the IMU simulation incorporates drifting. Hard impacts on the virtual robot will cause the IMU to drift, just like in real-world.
[Source Code | Video]
Building on the PhotonLib Camera Sim, this project features a fully realistic simulation of our custom AprilTag vision odometry.
We've also optimized the vision odometry by incorporating 6328's Log-Replay Technology, which allows for adjustments to our pose estimation strategy during replay, as detailed in this talk. We also offer a convenient feature that automatically generates the Advantage Scope Fix Camera Config, making it easy to set up and view the camera's perspective in Advantage Scope. Source Code.
-
- Generate TunerConstants.java using CTRE's Swerve Project Generator and drop it to
src/main/java/constants/
. - In vscode, use
Control
+H
to replace allprivate
tags inTunerConstants
withpublic
- Delete the last two lines
// TODO: the last two lines MUST be removed public static final CommandSwerveDrivetrain DriveTrain = new CommandSwerveDrivetrain(DrivetrainConstants, FrontLeft, FrontRight, BackLeft, BackRight);
- Edit a few variables in DriveTrainConstants.java
/** * numbers that needs to be changed to fit each robot * TODO: change these numbers to match your robot * */ public static final double WHEEL_COEFFICIENT_OF_FRICTION = 0.95, ROBOT_MASS_KG = 40; // with bumpers /** * TODO: change motor type to match your robot * */ public static final DCMotor DRIVE_MOTOR = DCMotor.getKrakenX60(1), STEER_MOTOR = DCMotor.getFalcon500(1);
- Try Driving Your Chassis On Field !!!
- Generate TunerConstants.java using CTRE's Swerve Project Generator and drop it to
-
β οΈ This project supports Rev hardware. However, it has not been tested on a physical Rev Chassis simply because we don't have one. If your team has an Rev Chassis and willing to test and debug our code, you are absolutely welcome and appreciated.- in RobotContainer.java line 100, change all
ModuleIOTalon
toModuleIOSpark
/* CTRE Chassis: */ drive = new SwerveDrive( new GyroIOPigeon2(TunerConstants.DrivetrainConstants), new ModuleIOTalon(TunerConstants.DrivetrainConstants, TunerConstants.FrontLeft, "FrontLeft"), new ModuleIOTalon(TunerConstants.DrivetrainConstants, TunerConstants.FrontRight, "FrontRight"), new ModuleIOTalon(TunerConstants.DrivetrainConstants, TunerConstants.BackLeft, "BackLeft"), new ModuleIOTalon(TunerConstants.DrivetrainConstants, TunerConstants.BackRight, "BackRight") ); /* REV Chassis */ drive = new SwerveDrive( SwerveDrive.DriveType.CTRE_ON_CANIVORE, new GyroIOPigeon2(TunerConstants.DrivetrainConstants), new ModuleIOSpark(0), new ModuleIOSpark(1), new ModuleIOSpark(2), new ModuleIOSpark(3) );
- in ModuleIOSpark.java, configure the motor IDs as well as the encoder offsets
- In DriveTrainConstants.java, replace all references to
TunerConstants
with actual data
- in RobotContainer.java line 100, change all
-
in Robot.java line 19, change
JAVA_SIM_MODE
toSIM
to run physics simulation or change it toREPLAY
to perform 6328's Log-Replay Technology.// Run physics simulation private static final Constants.RobotMode JAVA_SIM_MODE = Constants.RobotMode.SIM; // Run Log-Replay technology private static final Constants.RobotMode JAVA_SIM_MODE = Constants.RobotMode.REPLAY;
π‘ There is no need to change the mode to
REAL
when deploying to real robot as the program will automatically switch toREAL
mode if it is detected to be running on a real robotpublic static final Constants.RobotMode CURRENT_ROBOT_MODE = isReal() ? Constants.RobotMode.REAL : JAVA_SIM_MODE;
-
Run the WPILib Java Simulation and connect a gamepad to port 0
-
Open Advantage Scope and connect to simulator, add a new 3D Field Widget
-
Drag
AdvantageKit/RealOutputs/Field/Robot
to2D Poses
, make itRobot
(this is where your robot actually is in the simulation world)Drag
AdvantageKit/RealOutputs/Field/Robots
to2D Poses
, make itRed Ghost
(these are opponent robots)Drag
AdvantageKit/RealOutputs/Odometry/Trajectory
to2D Poses
, make itTrajectory
(this is the current activate path)Drag
AdvantageKit/RealOutputs/Odometry/RobotPosition
to2D Poses
, make itBlue Ghost
(this is where the odometry think the robot is) -
Open three Sendable Choosers in
SmartDashboard/FieldSimulation/
calledOpponentRobot1/2/3 Behavior
in the dashboard. SelectAuto Cycle
to let AI control these robots to do cycles, or selectJoystick Control Left/Right-Handed
to control them manually with another gamepad (port 1/2/3 for opponent robot 1/2/3)
π‘ The
subsystems/vision/apriltags
package is our team's implementation of an apriltag vision odometry, which is built on top of Photon Vision. We've implemented custom filtering mechanisms that enhance accuracy. We've also provide improved support for 6328's Log-Replay Technique.
β οΈ For teams that already have a reliable vision odometry solution in place, we recommend deleting this package and continuing to use the approach that best suits your needs.
On the vision coprocessors, please enable MultiTag and update the correct field layout according to the PhotonVision documentation.
![enabling multitag.webp](media/enabling multitag.webp)
Optionally, if you want to view the individual camera-to-target
transform calculated by SolvePNP
for each target during replay, please turn on Always Do Single-Target Estimation
. Note that this hurts performance a little.
In constants/VisionConstants.java
, put the infos about your camera here:
public static final List<PhotonCameraProperties> photonVisionCameras = List.of(
// declare all the photon cameras here
);
Now the vision system is good to go! Additionally, when a PhotonCameraProperties
instance is created, it will automatically print a line of JSON
code containing the position of the Fixed Camera.
You can use this output to easily configure a custom robot asset for Advantage Scope.
The vision odometry uses a technique called selective logging developed by Team 6328.
It only logs the raw inputs from the cameras and the final pose estimation result. If there's an issue with the result and you need to debug it, you must replay the code to observe the detailed filtering process and pinpoint the problem.
Open the log file using AdvantageScope. And change java sim mode to Replay
in Robot.java
. Run Simulate Java.
// Run Replay
private static final Constants.RobotMode JAVA_SIM_MODE = Constants.RobotMode.REPLAY;
Now Open the Replayed Log in AdvantageScope. The replayed log is named ORIGINAL-LOG-FILE-NAME_replayed.wpilog
Drag Field AdvantageKit/RealOutputs/AprilTags/Filtering/VisibleFieldTargets
to 3D Poses
, make it Vision Target
(these are the AprilTags on the field that are seen by one or more cameras).
Drag Field AdvantageKit/RealOutputs/AprilTags/Filtering/AprilTagsObservedPositions
to 3D Poses
, make it AprilTag 36h11
(this represents the observed positions of the tags on the field, assuming the current odometry pose is correct)
Drag Field AdvantageKit/RealOutputs/AprilTags/Filtering/ValidPoseEstimations
to 3D Poses
, make it Green Ghost
(these are the pose estimation results that the filtering mechanism has deemed valid)
Drag Field AdvantageKit/RealOutputs/AprilTags/Filtering/InvalidPoseEstimations
to 3D Poses
, make it Red Ghost
(these are the pose estimation results that the filtering mechanism has marked as invalid or noisy)
This is how it should look: (log file during WRC Beijing Off-season Event, Elimination 4)
You can see a lot of red results in the output, indicating that something is wrong.
Many AprilTags observed on the right-hand side of the robot appear to be floating in midair, likely due to an issue with the front-right camera.
To troubleshoot, we removed the FrontRightCam
from the camera configuration,
effectively disabling the camera for the replay.
Upon running the replay again, the red "ghost" errors disappeared.
It turns out the camera stand was bent.
This is a great example of how log-replay can quickly help you identify and resolve issues.
Since the vision odometry logs all the raw camera inputs, you will be able to change your pose estimation code and filtering mechanism during replay.
TODO: write the documents on how to fine tune the filtering mechanism using replay