Google Gemini, a multimodal AI by DeepMind, processes text, audio, images, and more. Gemini outperforms in AI benchmarks, is optimized for varied devices, and has been tested for safety and bias, adhering to responsible AI practices.
State-of-the-art solutions adopt the DETR-like framework, and mainly develop the complex decoder, e. g., regarding pose estimation as keypoint box detection and combining with human detection in ED-Pose, hierarchically predicting with pose decoder and joint (keypoint) decoder in PETR.
In this paper, we study the end-to-end multi-person pose estimation and present a simple yet effective transformer approach, named Group Pose. We simply regard �-keypoint pose estimation as predicting a set of �� keypoint positions, each from a keypoint query, as well as representing each pose with an instance query for scoring � pose predictions.
Motivated by the intuition that the interaction, among across-instance queries of different types, is not directly helpful, we make a simple modification to decoder self-attention. We replace single self-attention over all the �×(�+1) queries with two subsequent group self-attentions: (i) � within-instance self-attention, with each over � keypoint queries and one instance query, and (ii) (�+1) same-type across-instance self-attention, each over � queries of the same type. The resulting decoder removes the interaction among across-instance type-different queries, easing the optimization and thus improving the performance. Experimental results on MS COCO and CrowdPose show that our approach without human box supervision is superior to previous methods with complex decoders, and even is slightly better than ED-Pose that uses human box supervision.
Google Gemini, a multimodal AI by DeepMind, processes text, audio, images, and more. Gemini outperforms in AI benchmarks, is optimized for varied devices, and has been tested for safety and bias, adhering to responsible AI practices.
Cerelyze - Enabling engineers to rapidly reproduce scientific research
Video ReTalking, advanced real-world talking head video according to input audio, producing a high-quality
Then transplant it to the real world to solve complex problems
LongLLaMA is a large language model designed to handle very long text contexts, up to 256,000 tokens. It's based on OpenLLaMA and uses a technique called Focused Transformer (FoT) for training. The repository provides a smaller 3B version of LongLLaMA for free use. It can also be used as a replacement for LLaMA models with shorter contexts.
Large Language and Vision Assistant
Quick compare routes for nearby alternatives.
Compare Group Pose: A Simple Baseline for End-to-End Multi-person Pose Estimation with Free Google Gemini: the best largest and most capable AI model and jump into the preserved compare route.
Open compare route →Compare Group Pose: A Simple Baseline for End-to-End Multi-person Pose Estimation with Cerelyze-the Best AI Tools of paper to code and jump into the preserved compare route.
Open compare route →Compare Group Pose: A Simple Baseline for End-to-End Multi-person Pose Estimation with Video ReTalking-focuses on audio-based lip synchronization for talking head video editing and jump into the preserved compare route.
Open compare route →