|
--- |
|
license: apache-2.0 |
|
tags: |
|
- flight-planning |
|
- transformer |
|
- coordinate-prediction |
|
- sequence-to-sequence |
|
- count-classification |
|
--- |
|
# Flight Plan Coordinate Prediction Model (Seq2SeqCoordsTransformer) |
|
The encoder-decoder transformer model was trained for an AI flight planning project. Predicts normalized coordinates directly and waypoint count via classification. |
|
## Model Description |
|
Seq2SeqCoordsTransformer architecture using `torch.nn.Transformer`. Predicts normalized lat/lon coordinates autoregressively and waypoint count (0-10) via classification head on encoder output. |
|
* Embed Dim: 256, Heads: 8, Enc Layers: 4, Dec Layers: 4, Max Waypoints: 10 |
|
## Intended Use |
|
Research prototype. **Not for real-world navigation.** |
|
## Limitations |
|
Accuracy depends on data/tuning. Fixed max waypoints (10). Not certified. **Architecture differs significantly from previous versions in this repo.** |
|
## How to Use |
|
This requires loading the custom `Seq2SeqCoordsTransformer` class and weights. Generation requires autoregressive decoding and taking the argmax of the count logits. |
|
|
|
Read this article - https://medium.com/ai-simplified-in-plain-english/building-a-transformer-model-with-seq2seq-architecture-for-flight-planning-0bdd1fecaefe |
|
|
|
|
|
## Training Data |
|
Trained on `frankmorales2020/flight_plan_waypoints` - https://huggingface.co/datasets/frankmorales2020/flight_plan_waypoints. |
|
## Contact |
|
Frank Morales, BEng, MEng, SMIEEE (Boeing ATF) - https://www.linkedin.com/in/frank-morales1964/ |