Skip to content
@Ayn-Path

AynPath

Final Year Project (FYP) for the Bachelor of Computer Science program at IIUM, Malaysia.

AynPath: AI-Driven Smart Indoor Navigation System with AR for the Visually Impaired

AynPath

Overview

Navigating inside large buildings is often difficult for people with disabilities, especially those who are blind or have low vision. Most navigation tools today are designed for outdoor use and depend on GPS, which is unreliable indoors because of weak satellite signals. This makes it challenging to move through complex spaces such as campuses, shopping malls, or office buildings. Existing assistive apps usually focus on one specific feature, like detecting objects, reading text aloud, or guiding someone across a street. To get the full navigation experience, users often have to switch between several different apps, which can be inconvenient and overwhelming.

AynPath is my Final Year Project (FYP), developed to bring these needs together in a single application. By combining artificial intelligence and augmented reality, the system offers accurate indoor localization, detects obstacles commonly found in buildings, and provides real-time guidance. Directions are delivered through clear audio cues, gentle haptic feedback, and AR-based visual arrows. For this project, the implementation is focused on Block A, Level 1 of KICT IIUM, serving as the test environment for indoor navigation. The goal of AynPath is to give blind and visually impaired people more independence and confidence when moving through indoor environments, making navigation safer, smoother, and easier.

Diagram:

  1. Use Case Diagram

Indoor_Navigation_Flowchart

  • Scan Environment: Initiates the system by performing ORB feature detection to recognize virtual markers or detect the user’s location within the indoor environment using a custom VPS.

  • Select Destination: Allows the user to choose their intended destination within the predefined area of the KICT building.

  • AR-Based Navigation: Represents the primary navigation process by displaying virtual arrows as directional guidance toward the selected destination.

  • Receive Audio Guidance: Provides real-time auditory instructions during navigation to assist the user in following the correct path.

  • Get Obstacle Alerts: Notifies the user of any obstacles along the path to ensure safe and accurate navigation.

  1. Flowchart Diagram

The flowchart below illustrates the main processes in the Ayn-Path system, from scanning the environment to reaching the destination.

Indoor_Navigation_Flowchart

Repositories

Repository Description
aynpath-datasets Training and testing datasets for obstacle detection and indoor feature recognition. Includes annotated images, extracted ORB features, with the code for extracting them.
aynpath-app Flutter-based AR navigation app. Contains the full user interface and functional modules, integrating localization, obstacle detection, audio guidance, haptic feedback, and AR visual arrows.
aynpath-server Backend service for indoor localization and navigation. Provides APIs for VPS matching of the user’s current location using ORB features and image-based positioning.
aynpath-object_detection Object detection module powered by MobileNetV1 (TensorFlow Lite). Includes a pretrained model based on the COCO dataset, optimized for detecting common indoor obstacles.

Tools Used

  • Flutter
  • Unity
  • Python

Video Demonstration

Acknowledgements

Supervised by Dr. Nurazlin binti Zainal Azmi

Pinned Loading

  1. .github .github Public

Repositories

Showing 5 of 5 repositories

People

This organization has no public members. You must be a member to see who’s a part of this organization.

Top languages

Loading…

Most used topics

Loading…