Skip to content

Latest commit

 

History

History
66 lines (45 loc) · 3.05 KB

README.md

File metadata and controls

66 lines (45 loc) · 3.05 KB

ActiveLook® SPEECH

Description : What is said around you is translated & displayed in your ActiveLook® A/R glasses

The application can be found on GooglePlay : https://play.google.com/store/apps/details?id=com.speech.demo

License

Licensed under the Apache License, Version 2.0 (the “License”);
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
    http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an “AS IS” BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.

Requirements

You will need the following:

  • A pair of ActiveLook® glasses
  • Android Studio
  • An android device with BLE

Known supported Devices :

File to create and add at the root : .env

First, you need to add a file called '.env' at the source of the project. This file will contain only 2 lines :

ACTIVELOOK_SDK_TOKEN = ""
ACTIVELOOK_CFG_PASSWORD = 0xDEADBEEF

Main files to modify

The name of the app is defined in the strings.xml file.

  • app\src\main\res\layout\content_scrolling.xml
  • app\src\main\res\values\strings.xml
  • app\src\main\java\com\speech\demo\MainActivity.java

In order to get the best performances, the ActiveLookSDK directory should be the latest release from : https://github.com/ActiveLook/android-sdk

detailled description of this Android application

This 'speech-to-text' or voice recognition demo application is dedicated to deaf and hard of hearing people : it displays in your ActiveLook® glasses all of what is said around you, and what you are saying. If you choose another language, the translation will be shown.

The default language for the voice recognition is the one of your device, but you can change it to one of the 60 languages understood by the application from all over the world. The text can now be translated to one of the 60 languages are displayed in your connected glasses. You can also change the size of the text in your connected glasses depending on your convenience.

The app is based on GOOGLE-API for the speech recognition and GOOGLE-MLKit for the translation, so the app has their performance and limitation. For instance, it does not recognize what is said when you are watching TV.

This "ActiveLook® Speech" application connects to any Activelook® augmented reality glasses to display, live, and right in your field of vision, the key information you need to keep you always informed. The application will first pair via BTLE with your Activelook® smart glasses.