Skip to content

Latest commit

 

History

History
80 lines (68 loc) · 3.46 KB

INSTALLATION.md

File metadata and controls

80 lines (68 loc) · 3.46 KB

Preliminaries

  • Clone the repository

    git clone [email protected]:nagyrajmund/gesticulating_agent_unity.git
    cd gesticulating_agent_unity
    
    git submodule init
    git submodule update
    
  • Download Apache ActiveMQ 5 from this link and extract the

    • Extract the archive into the ActiveMQ folder
  • Download the Unity project:

Installation

Option 1: using docker (recommended)

  • Pull the docker image of the gesture generation model
    docker pull rajmundn/gesticulating_agent:gesturebot_blenderbot
    

Option 2: using anaconda (alternative)

  • Create a new conda environment and install the requirements (using the terminal on Linux or the Anaconda Prompt on Windows) by running the following commands from the root of the repository:
    # Create conda environment
    conda create --name gesturebot_blenderbot -y python=3.7
    conda activate gesturebot_blenderbot
    
    # Install gesture generation model
    cd gesticulator
    python install_script.py
    
    # Install Mozilla TTS
    cd gesticulator/interface/TTS_repo
    pip install -r requirements.txt
    python setup.py develop
    sudo apt-get install espeak
    
    # Download TTS model files
    gdown --id 1NFsfhH8W8AgcfJ-BsL8CYAwQfZ5k4T-n -O tts_model.pth.tar
    gdown --id 1IAROF3yy9qTK43vG_-R67y3Py9yYbD6t -O config.json
    gdown --id 1Ty5DZdOc0F7OTGj9oJThYbL5iVu_2G0K -O vocoder_model.pth.tar
    gdown --id 1Rd0R_nRCrbjEdpOwq6XwZAktvugiBvmu -O config_vocoder.json
    gdown --id 11oY3Tv0kQtxK_JPgxrfesa99maVXHNxU -O scale_stats_vocoder.npy
    
    cd ..
    

Running the project

Step 1: ActiveMQ

Start the ActiveMQ server by running ./bin/activemq start in a terminal (on Linux) or bin/activemq start in a command prompt (on Windows).

Step 2: Gesture generation model

  • Note that the gesture generation model will download around 10 GBs of data (for the language model) into the .word_vectors_cache folder when it's run for the first time. However, the 6,6 GB wiki.en.vec file can be removed after the first run.

Option 1: using docker

  • When the project is run for the first time, the docker container may be created by running the following command from the root of the repository:
    docker run -v $(pwd)/Unity/blenderbot_demo/gesturebot_Data:/workspace/gesticulator/interface/docker_volume --network host -ti --name gesturebot_blenderbot rajmundn/gesticulating_agent:gesturebot_blenderbot
    
  • After the container has been created, it can be ran with:
    docker start -ai gesturebot_blenderbot
    

Option 2: using anaconda

  • in the previously created conda environment, from the gesticulator folder, run:
    cd gesticulator/interface
    python main.py
    

Step 3: Running Unity

Run the Unity executable (gesturebot.x86_64 or gesturebot.exe in the Unity/blenderbot_demo/ folder)

Now you should be able to talk with the agent via the following ways:

  • Enter text in the input field and press Submit OR
  • Click Talk to start recording speech input, and Stop to stop recording OR
  • Press t-key once to start recording speech input, and t-key again to stop recording

and the agent should be moving when it replies.