Skip to content

Latest commit

 

History

History
71 lines (48 loc) · 2.06 KB

File metadata and controls

71 lines (48 loc) · 2.06 KB

Natural Language Interface

This is the code repository for the Android app developed by graduate students at Osnabrück University during the study projects supervised by Prof. Dr. Peter König:

  • Sensory Augmentation and Grasping Movements
  • Spatial Navigation supported by AI

Testing APK

Roadmap

	gitGraph LR:
		commit id: "Current state"
		commit id: "Workgroup release" tag: "v0.0.1"
		commit id: "Run example TensorFlow model"
		commit id: "Feedback round with experimental group"
		commit id: "Integrate feedback" tag: "v0.0.2"
		commit id: "Christmas break"
		commit id: "Object detection via any image input"
		commit id: "Polishing"
		commit id: "Optivist Presentation" tag: "v0.0.3"
		commit id: "Prepare for codebase transfer"
		commit id: "Location functionalities"
		commit id: "Improved GUI" tag: "v0.0.4"
		commit id: "Integrated LLM interface" tag: "v0.0.5"
Loading

Building the app locally

Requirements:

  • At least Java 17 JDK installed
  • Create a local.properties file at the root of this repository with the following content
MAPS_API_KEY=your_google_maps_api_key

You can request your keys via the Google Maps Platform.

Then build the application with

./gradlew build -x lint -x lintVitalRelease

on MacOS or Linux based operating systems.

The resulting installable application can be found under

app/build/outputs/apk/debug/app-debug.apk

which you can copy and install on your device.

Documentation

Running locally

  • Install Python
  • pip install -r docs/requirements.txt

Run the documentation server via

$ mkdocs serve

Integrations