Deep Transfer Learning for WiFi Localization
- URL: http://arxiv.org/abs/2103.05123v1
- Date: Mon, 8 Mar 2021 22:21:40 GMT
- Title: Deep Transfer Learning for WiFi Localization
- Authors: Peizheng Li, Han Cui, Aftab Khan, Usman Raza, Robert Piechocki, Angela
Doufexi, Tim Farnham
- Abstract summary: This paper studies a WiFi indoor localisation technique based on using a deep learning model and its transfer strategies.
We achieve a localisation accuracy of 46.55 cm in an ideal $(6.5m times 2.5m)$ office with no obstacles, 58.30 cm in an office with obstacles, and 102.8 cm in a sports hall.
- Score: 4.260395796577057
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper studies a WiFi indoor localisation technique based on using a deep
learning model and its transfer strategies. We take CSI packets collected via
the WiFi standard channel sounding as the training dataset and verify the CNN
model on the subsets collected in three experimental environments. We achieve a
localisation accuracy of 46.55 cm in an ideal $(6.5m \times 2.5m)$ office with
no obstacles, 58.30 cm in an office with obstacles, and 102.8 cm in a sports
hall $(40 \times 35m)$. Then, we evaluate the transfer ability of the proposed
model to different environments. The experimental results show that, for a
trained localisation model, feature extraction layers can be directly
transferred to other models and only the fully connected layers need to be
retrained to achieve the same baseline accuracy with non-transferred base
models. This can save 60% of the training parameters and reduce the training
time by more than half. Finally, an ablation study of the training dataset
shows that, in both office and sport hall scenarios, after reusing the feature
extraction layers of the base model, only 55% of the training data is required
to obtain the models' accuracy similar to the base models.
Related papers
- Transferable Post-training via Inverse Value Learning [83.75002867411263]
We propose modeling changes at the logits level during post-training using a separate neural network (i.e., the value network)
After training this network on a small base model using demonstrations, this network can be seamlessly integrated with other pre-trained models during inference.
We demonstrate that the resulting value network has broad transferability across pre-trained models of different parameter sizes.
arXiv Detail & Related papers (2024-10-28T13:48:43Z) - FINE: Factorizing Knowledge for Initialization of Variable-sized Diffusion Models [35.40065954148091]
FINE is a method based on the Learngene framework to initializing downstream networks leveraging pre-trained models.
It decomposes pre-trained knowledge into the product of matrices (i.e., $U$, $Sigma$, and $V$), where $U$ and $V$ are shared across network blocks as learngenes''
It consistently outperforms direct pre-training, particularly for smaller models, achieving state-of-the-art results across variable model sizes.
arXiv Detail & Related papers (2024-09-28T08:57:17Z) - Data Efficient Acoustic Scene Classification using Teacher-Informed Confusing Class Instruction [11.15868814062321]
Three systems are introduced to tackle training splits of different sizes.
For small training splits, we explored reducing the complexity of the provided baseline model by reducing the number of base channels.
For the larger training splits, we use FocusNet to provide confusing class information to an ensemble of multiple Patchout faSt Spectrogram Transformer (PaSST) models and baseline models trained on the original sampling rate of 44.1 kHz.
arXiv Detail & Related papers (2024-09-18T13:16:00Z) - DataComp-LM: In search of the next generation of training sets for language models [200.5293181577585]
DataComp for Language Models (DCLM) is a testbed for controlled dataset experiments with the goal of improving language models.
We provide a standardized corpus of 240T tokens extracted from Common Crawl, effective pretraining recipes based on the OpenLM framework, and a broad suite of 53 downstream evaluations.
Participants in the DCLM benchmark can experiment with data curation strategies such as deduplication, filtering, and data mixing at model scales ranging from 412M to 7B parameters.
arXiv Detail & Related papers (2024-06-17T17:42:57Z) - Towards Seamless Adaptation of Pre-trained Models for Visual Place Recognition [72.35438297011176]
We propose a novel method to realize seamless adaptation of pre-trained models for visual place recognition (VPR)
Specifically, to obtain both global and local features that focus on salient landmarks for discriminating places, we design a hybrid adaptation method.
Experimental results show that our method outperforms the state-of-the-art methods with less training data and training time.
arXiv Detail & Related papers (2024-02-22T12:55:01Z) - Boosting Visual-Language Models by Exploiting Hard Samples [126.35125029639168]
HELIP is a cost-effective strategy tailored to enhance the performance of existing CLIP models.
Our method allows for effortless integration with existing models' training pipelines.
On comprehensive benchmarks, HELIP consistently boosts existing models to achieve leading performance.
arXiv Detail & Related papers (2023-05-09T07:00:17Z) - $\Delta$-Patching: A Framework for Rapid Adaptation of Pre-trained
Convolutional Networks without Base Performance Loss [71.46601663956521]
Models pre-trained on large-scale datasets are often fine-tuned to support newer tasks and datasets that arrive over time.
We propose $Delta$-Patching for fine-tuning neural network models in an efficient manner, without the need to store model copies.
Our experiments show that $Delta$-Networks outperform earlier model patching work while only requiring a fraction of parameters to be trained.
arXiv Detail & Related papers (2023-03-26T16:39:44Z) - Train/Test-Time Adaptation with Retrieval [129.8579208970529]
We introduce Train/Test-Time Adaptation with Retrieval ($rm T3AR$), a method to adapt models both at train and test time.
$rm T3AR$ adapts a given model to the downstream task using refined pseudo-labels and a self-supervised contrastive objective function.
Thanks to the retrieval module, our method gives the user or service provider the possibility to improve model adaptation on the downstream task.
arXiv Detail & Related papers (2023-03-25T02:44:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.