Skip to content

ShreejanKumar/Indoor-Navigation-using-Deep-Learning

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

10 Commits
 
 
 
 
 
 

Repository files navigation

Indoor-Navigation-using-Deep-Learning

The project initially considered two approaches for assisting visually impaired individuals in navigation: explicit guidance through a 3D representation or autonomous determination of pathways based on obstacle locations. After surveying visually impaired participants, the latter approach was chosen. Using ViLT for visual question answering, we detected potential obstacles in images and segmented them with ClipSeg. Centroids of segmented objects were then computed to inform users of obstacle locations relative to their position. Depth estimation models helped prioritize obstacles, indicating the nearest and farthest objects. Finally, an LLM was used to generate instructions, which were delivered via text-to-speech based on survey insights.

image

Dataset Link - https://www.kaggle.com/datasets/dotran0101/daquar
ViLT - https://arxiv.org/abs/2102.03334
ClipSeg - https://huggingface.co/docs/transformers/en/model_doc/clipseg
GLPN - https://huggingface.co/docs/transformers/main/en/model_doc/glpn

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published