CaptionEmporium/conceptual-captions-cc12m-llavanext Viewer ⢠Updated Jun 30, 2024 ⢠11M ⢠630 ⢠20
Eagle: Exploring The Design Space for Multimodal LLMs with Mixture of Encoders Paper ⢠2408.15998 ⢠Published Aug 28, 2024 ⢠88
view post Post 5169 Real-time DEtection Transformer (RT-DETR) landed in transformers 𤩠with Apache 2.0 license šš models: PekingU š demo: merve/RT-DETR-tracking-cocoš paper: DETRs Beat YOLOs on Real-time Object Detection (2304.08069)š notebook: https://github.com/merveenoyan/example_notebooks/blob/main/RT_DETR_Notebook.ipynbYOLO models are known to be super fast for real-time computer vision, but they have a downside with being volatile to NMS š„²Transformer-based models on the other hand are computationally not as efficient š„²Isn't there something in between? Enter RT-DETR!The authors combined CNN backbone, multi-stage hybrid decoder (combining convs and attn) with a transformer decoder. In the paper, authors also claim one can adjust speed by changing decoder layers without retraining altogether. The authors find out that the model performs better in terms of speed and accuracy compared to the previous state-of-the-art. 𤩠š„ 12 12 š 4 4 + Reply