Software Guide15 min read

Conduct Vision Fear Conditioning Software Manual

This manual provides instructions for installing, configuring, and operating the ConductVision Fear Conditioning Software, which is used to conduct fear conditioning experiments with video tracking capabilities on rodents. The document covers software installation, experiment setup, stimulus protocol configuration, video tracking, hardware adapter integration, data export, and troubleshooting procedures.

Updated Feb 7, 2026
14 sections

Get This Product

ConductVision: AI Driven Video Tracking

ConductVision: AI Driven Video Tracking

Request a Quote

Get Quote
Fear Conditioning System

Fear Conditioning System

Request a Quote

Get Quote

Related Accessories

Also Available on Amazon

As an Amazon Associate, ConductScience earns from qualifying purchases.

ConductVision Software – Fear Conditioning

Contents

  1. Install and Run Application
    • 1.1 Software Components
    • 1.2 Machine Requirements
  2. Maze Types
  3. Experiment Configuration
    • 3.1 New Experiment
    • 3.2 Configure Stimulus Protocol
    • 3.3 Configure Video Tracking
  4. Video Service Connection
  5. Experiment Run
    • 5.1 Run Experiment
    • 5.2 Result Tables
    • 5.3 Video Replay
    • 5.4 Main Steps
  6. Conduct Adapter
    • 6.1 Conduct Adapter Software Installation
    • 6.2 Start ConductAdapterUI
    • 6.3 Execution Status on ConductAdapterUI
    • 6.4 Fear Conditioning Commands
    • 6.5 Manual Operations
    • 6.6 TTL
    • 6.7 Conduct Vision Simulator for Testing
  7. Multiple Cages
  8. Properties
  9. Animal
  10. Export Data
  11. Troubleshooting
    • 11.1 Log Display
    • 11.2 Database Files and Media Data
  12. License
  13. Camera Setup

1. Install and Run Application

1.1 Software Components

ConductVisionF\ConductVisionPy
Contains AI based tracking service executables and libraries that users would not need to interact with directly.

ConductVisionF\ConductVisionUI
The main executable is ConductVisionUI\bin\ConductVisionUI. Double click on the executable to launch the application.

appFearCondition\bin
Contains the ConductAdapter software that interfaces between the ConductVision video tracking system and Fear Conditioning Hardware system.

The application uses image buttons extensively. Tooltips appear when hovering over UI elements.

1.2 Machine Requirements

Since the application is a GPU-based AI application, it extensively uses Deep Learning Framework, GPU programming and AI libraries.

We recommend a desktop or laptop with NVIDIA RTX series like RTX 3090 or RTX 4060. Please contact our support for the recommendation of the latest GPU machine.

2. Maze Types

Click on the maze type field to view or select a maze type. You can switch to another maze without having to restart the application.

The maze types are defined in the rodent package the user purchased. More advanced packages are available.

A screenshot of a computer
A screenshot of a computer

3. Experiment Configuration

The application provides a user interface to configure and execute experiments (protocols) and view execution results.

3.1 New Experiment

To create an experiment protocol, click on "FC Protocols" on the left panel. A menu item "New" appears as shown below:

Click on "New" menu item, a "New Experiment" window appears as shown below. Fill in an experiment name and click on Save button.

A screenshot of a computer
A screenshot of a computer

Once the protocol is created, it appears in the folder "FC Protocols" as shown below:

A screenshot of a computer
A screenshot of a computer

An experiment can be deleted by right clicking on the experiment. The current active protocol is indicated in the left bottom of the screen.

3.2 Configure Stimulus Protocol

3.2.1 Protocol Stimulus Setup

A screenshot of a computer
A screenshot of a computer

The tab panel "FC Parameter" contains the stimulus types (light, tone, white noise, shock) and allows the user to add each type to the Protocol Action table. The parameters contain the stimulus timing dependencies as well.

The number of cages is configurable. The system supports up to 4 cages simultaneously tracking and analysis.

  • Copy icon allows the user to copy the protocol stimulus setup for one cage to another cage.

In the Protocol Action grid, the user can view, edit and test the Protocol Actions.

  • The folder button is used to load the MazeController from the ConductAdapterUI bin folder. The executable is used to send commands to ConductAdapterUI to control the hardware. The text field displays the executable file path.
  • The test button is used to send commands to ConductAdapterUI for manual testing.
  • The visualization button is used to visualize the stimulus actions in the FC Detection tab.

3.2.2 Test Commands

Before running the full tracking, it is helpful to run the commands to make sure communication between the ConductVision and ConductAdapterUI is properly established.

See ConductAdapter chapter for more details on the ConductAdapterUI side.

3.3 Configure Video Tracking

To configure video tracking parameters, double click on the experiment under the folder "FC Protocols" and go to "FC Detection" tab.

A screenshot of a computer
A screenshot of a computer

3.3.1 Tracking Parameters

  • Trial Durs(s) - represents the total video tracking duration in seconds
  • Stimulus Start(s) - represents the stimulus start time with a delay from the video tracking start
  • Cage Active – indicates if this cage is active in the tracking when there are multiple cages. If unchecked, the cage will not participate in the tracking
  • Head Device – indicate if the animal has an optogenetic device mounted on the head with wires. If there is no head device, a motion pixel change algorithm is used; otherwise, we use AI-based pose estimate model to detect the animal motion that filters out the wire movements
  • Source – indicate the video source is from a video file or real-time camera tracking
  • Video Source: Video file - to track a video file, the user needs to load a video file
  • Video Source: camera – this is to track rodent in real-time
  • Video File Path – if the video source is file, the video file needs to be loaded
  • Preview – in live tracking mode, the user can preview the maze; stop preview will produce a thumbnail that can be used for drawing the tracking areas
  • Draw cage – Draw cage bound rectangle as a tracking area
  • Camera ID – in case of multiple cages, camera ID needs to be assigned to each cage properly. Connect one camera to the computer, the cameraId field will have one camera ID and you can assign it to the cage. Then connect one more camera to the computer, the field will have two camera IDs, assign the second camera ID to the second cage. For example, if you have four cages, continue this process to assign the four cameras to four cages so that each cage has a unique camera ID.

3.3.2 Freeze Score and Stimulus Visualization

A screenshot of a computer
A screenshot of a computer

Freeze Score (green line in the chart) - cumulative freeze frames / total frames

Stimulus (blue and red line in the chart) - when the stimuli occur

Marker (purple line in the chart) – moves every second along with stimuli

Freeze circle – turns yellow when the current frame is a freeze frame and turns off when the frame is not a freeze frame. Turns red when the freezing frames become a freezing episode (the consecutive freeze frame duration exceeds the freezing criteria).

3.3.3 Freezing Episode

A freezing episode occurs if the animal is motionless for a duration equal to or greater than the parameter "Freeze Min Duration in Milliseconds."

A screenshot of a computer
A screenshot of a computer

4. Video Service Connection

To use AI based video tracking service, the user needs to start the service stream. When the service stream is established, the status field shows text "Started" and color becomes green. If there is a need to stop the stream because of an exception, click on the stop button and then start again.

When the application starts, it is in the stop state.

The stream uses local machine port. It is usually available for users. If the port is blocked, please contact your system administrator to open the port.

A screenshot of a computer
A screenshot of a computer

5. Experiment Run

To execute an experiment, select the protocol under "Protocols" on the left navigation panel and double click on it. Make sure the service stream is connected.

5.1 Run Experiment

Enter a name for the experiment run and click Start button. This will set system in the running state.

A screenshot of a computer
A screenshot of a computer

When the start button is clicked, the state button becomes Stop. The user can click on a Stop button to stop the session run.

The experiment will stop when the total trial duration is reached, or the user stops it.

5.2 Result Tables

Select the result from the left "Experiment Result." The results are shown on the Result tab. There are four grids.

A screenshot of a computer
A screenshot of a computer

5.2.1 Track Details

The table shows the details of each frame:

  • Seg Pixels – represents the pixel change from current frame compared to the previous frame
  • Freeze Score – is a metric used to quantify the degree of freezing behavior in animals during behavioral studies. It reflects the proportion of freezing (motionlessness) frames
  • IsFreezing – indicates if the frame is a freezing frame. A frame is a freezing frame if the pixel change is less than the predefined small number configured in the property. This parameter is not zero to account for noise

The Freeze Score column and IsFreeze column are not calculated during the tracking but will be calculated when the tracking session is completed.

5.2.2 Action Details

The table shows the stimulus action details.

5.2.3 Freeze Analysis

A screenshot of a computer
A screenshot of a computer

The table shows the freezing episodes and stimulus actions. The synchronized timestamp helps to analyze how an animal's freezing behavior aligns with the presentation of specific stimuli.

Each freeze episode row shows the freeze start time, end time, duration and corresponding sequence numbers in the tracking detail table above.

FreezeScoreMean is the mean value of the freeze scores for the entire time range [FromStart, end]. The value is in the column Value of the row.

FreezeScoreStdev is the standard deviation of the freeze scores for the entire time range [FromStart, end]. The value is in the column Value of the row.

The user can choose the time ranges and Freeze Min Duration to recalculate and analyze the freeze episodes.

5.2.4 Freeze Chart

A screenshot of a computer
A screenshot of a computer

The visualization of freeze episodes, stimuli and freeze scores based on the Freeze Analysis result. The display time range is default to the relevant time period from the Freeze Analysis table.

5.2.5 Reset Freezing Parameters

A screenshot of a computer
A screenshot of a computer

If the results generated by the current freezing parameters are not optimal, the user can adjust the parameters to generate new results.

Fear Conditioning Parameters:

  • SegFreezePixels: Defines the sensitivity threshold (in pixels) for detecting freezing behavior. Frames where the movement is less than this value are considered potential freezing frames.
  • RollingWindowFrames: Specifies the number of consecutive frames used for calculating smoothed motion data. A larger window produces smoother results but may reduce responsiveness to short movements.

5.3 Video Replay

A screenshot of a computer
A screenshot of a computer

Replay Freeze Episode:
Click on an episode and then click Replay Freeze Episode. The selected freezing episodes will be displayed in the video replay panel.

Replay Frame to Frame:
Allows the user to define a specific frame range for replay.

The user can replay the tracking video at a slow speed to observe the animal behavior.

5.4 Main Steps

5.4.1 Hardware and ConductAdapterUI

  • Step 1: Connect hardware components and connect the USB cable from the computer to the control box.
  • Step 2: Start hardware control box and connect to the comport.
  • Step 3: Start ConductAdapter and test the hardware functionality using manual commands from ConductAdapterUI.

5.4.2 ConductVision

  • Step 4: Run ConductVision software and set up a protocol with the parameters. The parameters include video parameters and protocol stimulus parameters.
  • Step 5: Run the protocol. Make sure the commands are properly sent to the ConductAdapterUI.
  • Step 6: Review the result.

6. Conduct Adapter

ConductVision communicates with ConductAdapter through commands. The adapter controls the execution of the Fear Conditioning Hardware.

6.1 Conduct Adapter Software Installation

Maze Engineers software ConductAdapter is a light-weight adapter for ConductVision to control the execution of the Fear Conditioning Hardware.

The executable ConductAdapterUI is a light-weight Windows UI application that needs to start before experiments. The application accepts commands from ConductVision and controls the execution of the commands.

6.2 Start ConductAdapterUI

When the micro-controller is connected to your computer, a port is created as shown on the Device Manager:

A screenshot of a computer
A screenshot of a computer

If the device is not shown on the Ports section, it will appear in a different section. This is because the computer could not find the driver for the hardware device. In this case, you will need to install the driver (small disc in the box of UNITEK) which is included in shipping. A digital version of the USB driver is included in shipment so that you would not need a CD reader.

For Windows 11, the driver install appears to be unnecessary.

Step 3: Start ConductAdapterUI

Double click on ConductAdapterUI, the main window appears as below:

A screenshot of a computer
A screenshot of a computer

Step 4: Connect to comport

Once your computer recognizes the hardware, choose the comport that matches the one on the Device Manager, and then click on Connect button.

A screenshot of a computer
A screenshot of a computer

6.3 Execution Status on ConductAdapterUI

When the experiment runs, the External Command Action is invoked, and the command is sent to ConductAdapterUI. On the left upper panel of ConductAdapterUI, the command from ConductVision is logged. On the left lower panel of ConductAdapterUI, the command execution status is logged. ConductAdapterUI can be minimized if you don't need to monitor the status.

The right main panel of ConductAdapterUI provides functionality for users to send manual commands to the automated maze for testing purposes.

6.4 Fear Conditioning Commands

6.4.1 Command Parameters

Command parameters for ConductVision video tracking. They are automatically built in the ConductVision protocol panel. They are listed here for information purposes so you can see the command flow from ConductVision to the ConductAdapter.

TypeMessage TypeStart command exampleStopParameters
Frequency tonet=1c=1,t=1,v=50,d=5,f=1000c=1,t=1,v=0c: chamber 1, 2, 3 or 4
t: message types:
1: frequency tone
2: light
5: shock
6: white noise
8: StartTrial (for TTL signal only)
9: EndTrial (for TTL signal only)

v: volume 0-100
d: duration 0-180
f: frequency 100-18000
Lightt=2c=1,t=2,v=30,d=5c=1,t=2,v=0
Shockt=5c=1,t=5,v=1,d=5,m=0.4c=1,t=5,v=0
White noiset=6c=1,t=6,v=50,d=10c=1,t=6,v=0
StartTrialt=8c=1,t=8
EndTrialt=9c=1,t=9

6.5 Manual Operations

For testing purposes, the software allows the user to send commands (connection has to be established).

6.6 TTL

By default, a BNC TTL connector is equipped. The user can send the StartTrial and EndTrial signals to the external devices. The TTL connection is between the control box and the external device(s).

A screenshot of a computer
A screenshot of a computer

6.7 Conduct Vision Simulator for Testing

For testing purposes, the executable VideoSimulator can be invoked to generate commands to ConductAdapterUI software.

A screenshot of a video simulator
A screenshot of a video simulator

7. Multiple Cages

The number of cages is defined in the ConductVisionUI configuration file.

8. Properties

A screenshot of a computer
A screenshot of a computer

Fear Conditioning Parameters:

  • SegFreezePixels: Defines the sensitivity threshold (in pixels) for detecting freezing behavior. Frames where the movement is less than this value are considered potential freezing frames.
  • RollingWindowFrames: Specifies the number of consecutive frames used for calculating smoothed motion data. A larger window produces smoother results but may reduce responsiveness to short movements.

9. Animal

The tab provides a place for the user to enter animal information. The animal information is then applied to the results.

A screenshot of a computer
A screenshot of a computer

10. Export Data

A screenshot of a computer
A screenshot of a computer

  • Select a result
  • Click the Export button
  • The output file format is .csv

11. Troubleshooting

11.1 Log Display

  • The log content can be displayed on the screen.
  • Each time ConductVisionUI starts, it creates a new log file.

11.2 Database Files and Media Data

Two folders of database and media are generated during the operation.

For ConductVision support to replicate your tracking operation, the following data are needed:

  1. Your video file if the source is Video File
  2. Log files

12. License

Demo version: The demo version is valid for one month from the download and no license is required.

For us to generate a license, you will need to provide us with two machine names where Conduct Vision software is intended to run.

There are two ways to find the computer name.

Method 1: From Control Panel. It works for Windows 10 and Windows 11.

A screenshot of a computer program
A screenshot of a computer program

Method 2: For Windows 11, from Start button → System → About.

A screenshot of a chat
A screenshot of a chat

13. Camera Setup

If one of our provided cameras is used, no drivers or software is required for the camera to function.

  1. Disable any other cameras on the computer under "Device Manager"

A white screen with a blue background
A white screen with a blue background

  1. Plug the camera into the computer's USB port. You should hear a noise to indicate the computer has recognized the camera.
  2. Open the computer's camera app to confirm the camera's connection and adjust the camera's view.
  3. Either record subject trials in advance using your computer's camera app or open the ConductVision app for live streaming.

Was this article helpful?

Related Articles

Conduct Vision Software Manual

ConductVision Software Manual Contents 1\. Install and Run Application 7

ConductVision Specification

ConductVision is an AI-based animal behavior tracking software that operates at 30+ frames per second and uses machine learning to detect multiple body points for tracking various animal subjects including mice, rats, ferrets, zebrafish, and Drosophila. This specification document outlines the software's key features, advantages, and hardware requirements for computer and camera systems.

Camera Manual

This manual provides instructions for the Conduct Vision Camera, including connection to a PC, use of compatible video recording software, mounting options via a camera holder, and operation through the built-in Camera app on Microsoft Windows devices.

Camera Specification

This document specifies the technical parameters for two Conduct Vision infrared HD camera models: a Standard model with 2-megapixel resolution and 30 fps capability, and an Advanced model with 5-megapixel resolution and 60 fps capability. Both cameras feature CMOS sensors, multiple field-of-view options, and MJPG format output via USB interfaces.

ConductVision Catwalk Specification

This specification document describes the ConductVision Catwalk system, a gait analysis platform consisting of a glass walkway with LED illumination and high-speed camera that captures paw contact images, paired with AI-powered software that automatically detects and classifies paw contacts to measure spatial, temporal, kinetic, and coordination gait parameters in rodents.

Need more help?

Our support team is here to assist you