Skip to content

Run your AI and CV algorithms in meetings such as Zoom, Meets or Teams! 🚀

License

Notifications You must be signed in to change notification settings

nengelmann/MeetingCam

Repository files navigation

image

MeetingCam

✨ Special effects for not so special meetings. 👀


Run your AI and CV algorithms in online meetings such as Zoom, Meets or Teams! 🚀

This Project is one of the Popular Vote Finalists of the OpenCV AI Competition 2023. 🎉


CoverVideo2.mp4

First-person face detection prints your face detection and name in your webcam stream.
More plugins are listed further down the readme.

Intro video

Intro Video

Installation

  1. Clone this repo

    git clone https://github.com/nengelmann/MeetingCam.git && cd MeetingCam
  2. Install virtualenv

  3. Create a virtual python3.10 environment

    virtualenv -p /usr/bin/python3.10 .venv && source .venv/bin/activate
  4. Install the dependencies

    python -m pip install -r requirements.txt
  5. Setup pyvirtualcam

    sudo apt update
    sudo apt install v4l2loopback-dkms
    sudo apt install v4l-utils

Usage

First add a camera device to MeetingCam

To show a modified camera stream in online meetings it is necessary to create a virtual camera for each real camera you want to use.

  1. Make sure your webcam is connected.

  2. Activate the virtual environment

    source .venv/bin/activate
  3. Run the main.py to see general commands and plugin options.

    python ./src/meetingcam/main.py
  4. List available cameras

    python src/meetingcam/main.py list-devices
  5. Get command to add a camera device

    python src/meetingcam/main.py add-devices

    Then copy and execute the command of the camera you want to use.

  6. You can now see that the camera of your choice has a virtual counterpart.

    python src/meetingcam/main.py list-devices

Run a plugin

The general command to run a plugin is:

python src/meetingcam/main.py PLUGIN_NAME [OPTIONS] DEVICE_PATH

e.g. the face detection plugin:

python src/meetingcam/main.py face-detector --name yourname /dev/video0

Options and usage for each plugin are documented in the plugins help function
and in it's readme (link below 👇)

Available plugins


Detection of first persons face and imprint of bounding box with name.


Runs roboflow object-detection and instance-segmentation models.


Runs a Yolov5 model trained on COCO. Computation on a depthai device.

Custom plugins

You can create custom plugins from a predefined template to kickstart your project. 🚀

  1. Create plugin from template:

    python src/meetingcam/main.py create-plugin

    Alternative:

    python src/meetingcam/main.py create-plugin --name your_plugin --short-description "My awesome plugin"  --description "Plugin which runs some custom AI and CV algorithms"

    Make sure to add quotes " to sentences with spaces otherwise it will be interpreted as an extra argument.

  2. Run the template to ensure the creation worked out, see Run a plugin

    python src/meetingcam/main.py your_plugin /dev/your_device


    You'll see an imprint if you access the camera in a meeting tool.

  3. Go to src/meetingcam/plugins/ and you see your plugin as a directory which contains a plugin.py file.
    This file can be customized to run your CV and AI on the webcam stream!
    It boils down to a CustomPlugin class where you can initialize and process the images as well as a typer main function where you need to adapt custom arguments.

    • CustomPlugin class:
      This class needs to have an init and process function, for initialization and processing of a camera images respectively.

      class YourPluginClass(PluginBase):
      
         def __init__(self, your_arg) -> None:
            super().__init__()
            # some more custom init, e.g. your AI model
            self.your_arg = your_arg
            self.model = YourModel()
      
         def process(
            self,
            image: NDArray[Any],
            detection: Any,
            keyhandler: Type[KeyHandler],
         ) -> NDArray[Any]:
            # custom image processing
            image = self.model(image, your_arg)
            cv2.putText(image,"That's printed in every frame!", (10, image.shape[0]//2), cv2.FONT_HERSHEY_SIMPLEX, 1, 255)
            return image
    • Last step is to adapt your plugins entry point (typer app), if you have some arguments you want to pass via CLI.

      @plugin_app.callback(rich_help_panel="Plugin-Commands")
      def main(
         device_path: DevicePath = DevicePathWebcam,
         your_arg: Optional[str] = typer.Option(
            default=None, help="Your custom argument."
         ),
      ):
         # instantiate your plugin
         plugin = YourPluginClass(your_arg)
         # instantiate runner with plugin and device path
         runner = Runner(plugin, device_path)
         # run
         runner.run()

Hotkey triggers

There are two default triggers build in. They allow you to switch the image color channel from BGR to RGB and to mirror the video stream. The main purpose is, that if your camera is outputting an BGR stream by default you would look bluish, you can switch without modifying the code. Also the mirroring is helpful in some online meeting tools.

<Ctrl>+<Alt>+r -> Switch color channels (RGB<>BGR).
<Ctrl>+<Alt>+m -> Mirror the video stream.

Plugins can use additional triggers within their processing function. This is useful to hide some imprints or run the model inference when needed.

Custom hotkey triggers can be defined in the plugins main class initialization with by specifying a hotkeys list with Hotkey(key_combination, variable_name, is_enabled, description).

self.hotkeys = [
            Hotkey("<Ctrl>+<Alt>+z", "z_trigger", True, "toggle something"),
        ]

This trigger can be accessed during runtime in the plugins processing function as follows.

# If z_trigger <Ctrl>+<Alt>+z is True, print in additional text
if keyhandler.z_trigger:
   image = cv2.putText(
                image,
                "Toggle this text with <Ctrl>+<Alt>+z",
                (100, 100)
   )

Have a look at the available example plugins for more information or create your own.

Further information

Restrictions

Resolution
In general most meeting tools have the restriction on camera/video resolution, mostly 720p (1270x720 pixels). Make sure to not send a virtual video stream with higher resolution. If it breaks due to resolution you can select the virtual camera in the meeting tool but it will either show "Camera failed" or a black video stream.

Chrome/Chromium Browser
Chrome/Chromium might need to be ran with exclusive_caps=1 in the sudo modprobe v4l2loopback command which is not (yet) supported by MeetingCam.

Tool compatibility

Meets ✅

Works on Firefox.

Zoom ✅

Works on Firefox.

Teams 🚧

There are several Problems running MeetingCam on Linux with Teams.
Teams, at least on Linux is just running on Chromium, Edge and it's Progressive Web App (PWA). Therefore we need to run the sudo modprobe v4l2loopback command with exclusive_caps=1 argument, which makes troubles down the road. For some cameras this might work but is likely to fail. Even with the exclusive_caps=1 argument it is not reliably working.
A workaround is described here. On Firefox Version 117.0.1 (snap install) it was working with the overwrite Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/79.0.3945.74 Safari/537.36 Edg/79.0.309.43. This overwrite might cause unwanted behaviour (including Google Meets calls are failing). So it is advisable to not add the overwrite in your main firefox profile. However you can run firefox -p and create a new dedicated profile and apply the above changes. Then you can open it with firefox -p or firefox -P PROFILE_NAME and run it isolated.

Camera setup info

More information on how to setup virtual cameras can be found here

Development and modifications

You can customize this repo for your needs, you can also write your own AI-Plugin for running your models on Zoom, Teams or Meets.
More information about that in the section Custom plugins section and in DEVELOP.md Templates and documentation on how to run your custom AI will follow.

Usage with depthai camera

Run OAK devices as webcam (uvc)

Run OAK devices as uvc before normal usage of MeetingCam.

 source .venv/bin/activate
 python ./tools/oak_as_uvc.py

Run OAK devices with on device compute

Follow the steps in the Usage section normally, just append a --type depthai flag to the list-devices and add-devices command and use a plugin which is of type depthai.

 python src/meetingcam/main.py list-devices --type depthai
 python src/meetingcam/main.py add-devices --type depthai

Usage with Roboflow

For roboflow it is currently needed to run a roboflow inference server in a separate terminal.

You can run it e.g. on CPU with the following command:

docker run --net=host roboflow/roboflow-inference-server-cpu:latest

See here for more options.