VScan
A visual perception layer for the blind
新版本 0.2.3
- There is now a standalone editor for entering the system prompt and user prompt. This editor has a large text field, which should make it easy to work with long and complex prompts.
- Various UI improvements and bug fixes.
- Various UI improvements and bug fixes.
This is a little project of mine aiming to research how vision LLMs could help out blind people on travel and in their every-day life by substituting eyesight for various visual tasks. VScan turns your smartphone's camera into a device for visual perception. You can define various optical cognitive functions, like looking for objects, signs, evaluating a scene or simply mediating visual impressions. You can afterwards use these functions on the camera view, just like a sighted person would use their eyes to achieve a specific goal in the physical world.
Each cognitive tool consists of two major parts:
- The camera to be used - front / back, as well as camera parameters - resolution, flashlight etc.
- The prompts used for LLM processing. LLM is the bridge between raw pixel data and your interpretation of it, and in the user/system prompt, you can specify what exactly are you interested in for the particular function and how should it be communicated, as well as the LLM model that should be used.
Camera input in combination with an LLM processing prompt forms a cognitive function, which can be used to serve various visual tasks.
VScan is open-source software. Visit the project's official repository to learn more about its background, motivation, specific usage examples and setup instructions.
版本
雖然在下方可選擇下載 APK 檔案,但要留意這樣的安裝方式將不會收到更新通知,是一種較不安全的下載方法。建議您先安裝 F-Droid 用戶端使用。
下載 F-Droid-
arm64-v8aarmeabi-v7ax86x86_64此版本需要 Android 7.0 或更高的版本。
此套件包由原開發者構建和簽署,並保證與此原始碼 Tarball 保持一致。



