Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Include a new function to improve quality of training data using BERT. #1256

Open
gilbertocamara opened this issue Dec 16, 2024 · 0 comments

Comments

@gilbertocamara
Copy link
Contributor

Selecting high-quality training samples is crucial for enhancing the accuracy of land classification using remote sensing data. A significant challenge in this area is the limited availability of large datasets that contain good quality training samples. As a result, the deep learning community has explored various techniques to maximize the potential of small training datasets. One such method is SITS-BERT, which stands for "Bidirectional Encoder Representations from Transformers" applied to satellite image time series.

BERT, a technique developed in the first generation of large language models (LLMs), employs a combination of unsupervised pretraining and supervised fine-tuning. It utilises a "Masked Language Model" as its pretext task. The model predicts these missing tokens based on the surrounding context after this task randomly masks certain tokens in a sentence. This approach allows BERT to learn word relationships from a plain text corpus.

When applied to satellite image time series (SITS), the BERT method begins by training a deep learning model to recover missing observations within the time series. A second phase then refines this model using additional training samples. It is worth investigating whether the SITS-BERT approach can enhance the quality of the models used for land classification.

Reference:

Yuan, Yuan, and Lei Lin. 2021. ‘Self-Supervised Pretraining of Transformers for Satellite Image Time Series Classification’. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 14:474–87. https://doi.org/10.1109/JSTARS.2020.3036602.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
Status: To do
Development

No branches or pull requests

4 participants