--- title: Video Human Fall Detection With TimeSformer emoji: 🌍 colorFrom: blue colorTo: pink sdk: gradio sdk_version: 5.25.0 app_file: app.py pinned: false license: apache-2.0 short_description: Fall detector with TimeSformer --- # Video Human Detection Demo using TimeSformer This is a Hugging Face Spaces demo project that uses TimeSformer – a video transformer model – for video-based human detection (action recognition). In this demo, we use the pre-trained model `microsoft/timesformer-base-finetuned-k400` from Hugging Face, which has been fine‑tuned on the Kinetics‑400 dataset. The model is capable of classifying a video into one of 400 human action categories. ## Overview - **Model:** We use a TimeSformer model (`microsoft/timesformer-base-finetuned-k400`) to classify video clips. - **Feature Extractor:** The demo employs the Hugging Face `AutoFeatureExtractor` for video to process and prepare video frames. - **Inference:** The model outputs a set of predicted action labels with scores. These predictions help detect human actions in the video. - **Interface:** Built with Gradio, the demo lets the user upload a video file. The application extracts frames from the video, processes them with the model, and displays the top action predictions. ## Setup and Deployment 1. **Requirements:** See `requirements.txt` for the list of required packages. 2. **Run Locally:** You can run the demo locally using: ```bash python app.py ``` 3. **Deploy on Hugging Face Spaces:** Simply push these files to a new repository under HF Spaces. The app is designed to run with ZeroGPU if available and it is fully compatible with CPU-only environments. ## Notes - **Video Preprocessing:** The demo extracts frames using OpenCV and passes them to the feature extractor. The number of frames and the resolution are set to default values that can be adjusted. - **Model Performance:** TimeSformer is computationally heavy – for real-time use, consider using a smaller or distilled variant, or reduce the number of frames processed. - **ZeroGPU Support:** The app uses the `@spaces.GPU` decorator (from the HF Spaces ZeroGPU environment) if available; otherwise, it will run on CPU. Enjoy testing human detection in videos with this demo!