BlazePose GHUM 3D: Transfer Learning

How can we use the pre-trained BlazePose GHUM 3D model to apply transfer learning to the addition of custom set of images because I want to include people using wheelchairs and people using walkers and crutches as well?

Using Tensorflow.js, I’d like to also track the different joints of a simulated 3D human
character displayed on a webpage. As of now, BlazePose GHUM models will not work as they were
only trained on humans.

How can we use transfer learning in this context so that I can train 3D pose detection for virtual human characters displayed on the web page using BlazePose GHUM?

Thanks in advance!

Hi, the pose models are not suitable for transfer learning. If you have a specific data set and use case, you can train your own model.

You can find some more information on this thread: https://tensorflow-prod.ospodiscourse.com/t/open-source-movenet/3829/3?u=lgusm

But it’s about MoveNet and not BlazePose

Maybe you can somehow combine two models - one for pose detection and one for object detection - to solve your problem? If the bounding box for specific objects (wheelchairs, etc.) overlaps with specific key points of a human body identified by the pose model, you can be more or less sure what the image contains.

Given monocular RGB images of specifically individuals in wheelchairs, how can we reconstruct the 3D pose of humans?

Are there any example Python code so I don’t have to build it from scratch for training my own custom images?

This will be empowering for people with disabilities worldwide in the healthcare domain and increase accessibility in many aspects.

Unfortunately, you will hardly find code examples that you can use out-of-the-box. What I suggested was that you probably could take a pretrained object detection model, which is suited for transfer learning, and tune it to detect specialised equipment like wheelchairs and crutches.
Then you can take an image and get two predictions from two models: pose detection model will give you exact coordinates of the key points of a person, and fine-tuned object detection model will say if the image also contains the equipment and where it is located. And you’ll be able to identify positions of a person and equipment relative to each other.
Here is a list of object detection models that could be tuned: TensorFlow Hub
Hope that helps.

2 Likes