{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"VL-GPT","owner":"AILab-CVC","isFork":false,"description":"VL-GPT: A Generative Pre-trained Transformer for Vision and Language Understanding and Generation","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":2,"starsCount":84,"forksCount":2,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,1,3,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-12T05:25:46.171Z"}},{"type":"Public","name":"CV-VAE","owner":"AILab-CVC","isFork":false,"description":"CV-VAE: A Compatible Video VAE for Latent Generative Video Models","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":12,"starsCount":206,"forksCount":6,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,7,3,0,0,0,0,0,0,0,0,0,0,0,1,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-02T02:18:52.668Z"}},{"type":"Public","name":"YOLO-World","owner":"AILab-CVC","isFork":false,"description":"[CVPR 2024] Real-Time Open-Vocabulary Object Detection","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":6,"issueCount":274,"starsCount":4334,"forksCount":418,"license":"GNU General Public License v3.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,5,15,15,20,11,13,2,14,8,2,11,9,2,8,1,1,12,0,1,4,0,0,0,2,1,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-30T14:59:19.699Z"}},{"type":"Public","name":"HiFi-123","owner":"AILab-CVC","isFork":false,"description":"[ECCV 2024] HiFi-123: Towards High-fidelity One Image to 3D Content Generation","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":57,"forksCount":0,"license":null,"participation":[0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,1,2,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-12T09:42:35.515Z"}},{"type":"Public","name":"SEED-X","owner":"AILab-CVC","isFork":false,"description":"Multimodal Models in Real World","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":17,"starsCount":372,"forksCount":17,"license":"Other","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,9,0,0,4,8,14,2,1,2,0,0,0,1,0,0,2,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-12T09:06:56.500Z"}},{"type":"Public","name":"FreeNoise","owner":"AILab-CVC","isFork":false,"description":"[ICLR 2024] Code for FreeNoise based on VideoCrafter","allTopics":["generative-model","diffusion","aigc","video-diffusion-model"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":6,"starsCount":365,"forksCount":24,"license":"Apache License 2.0","participation":[0,0,0,0,0,27,10,2,2,1,2,0,0,0,0,0,0,0,4,4,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-11T13:19:52.334Z"}},{"type":"Public","name":"SEED-Bench","owner":"AILab-CVC","isFork":false,"description":"(CVPR2024)A benchmark for evaluating Multimodal LLMs using multiple-choice questions.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":19,"starsCount":303,"forksCount":9,"license":"Other","participation":[1,0,0,0,0,0,0,0,0,0,16,14,1,10,0,0,0,0,0,1,0,0,0,1,0,1,0,0,0,0,0,1,3,0,1,0,1,7,0,0,1,0,0,3,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-11T06:18:32.288Z"}},{"type":"Public","name":"AILab-CVC.github.io","owner":"AILab-CVC","isFork":false,"description":"Homepage of Tencent AI Lab CVC.","allTopics":[],"primaryLanguage":{"name":"HTML","color":"#e34c26"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":[0,0,0,3,4,2,6,3,1,0,0,0,0,3,3,0,0,1,2,1,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,2,0,0,0,0,1,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-10T08:35:17.112Z"}},{"type":"Public","name":"VideoCrafter","owner":"AILab-CVC","isFork":false,"description":"VideoCrafter2: Overcoming Data Limitations for High-Quality Video Diffusion Models","allTopics":["video-generation","image-to-video","text-to-video"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":3,"issueCount":61,"starsCount":4461,"forksCount":331,"license":"Other","participation":[0,0,0,1,2,2,3,7,0,0,1,1,0,0,0,0,0,0,2,2,4,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-10T08:30:45.710Z"}},{"type":"Public","name":"SEED","owner":"AILab-CVC","isFork":false,"description":"Official implementation of SEED-LLaMA (ICLR 2024).","allTopics":["multimodal","vision-language","foundation-model"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":32,"starsCount":556,"forksCount":31,"license":"Other","participation":[0,1,8,4,5,23,0,6,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,5,3,11,3,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-11T12:46:51.189Z"}},{"type":"Public","name":"TaleCrafter","owner":"AILab-CVC","isFork":false,"description":"[SIGGRAPH Asia 2023] An interactive story visualization tool that support multiple characters","allTopics":["storytelling","storycreation","siggraph-asia-2023","siggprah-asia"],"primaryLanguage":null,"pullRequestCount":0,"issueCount":5,"starsCount":250,"forksCount":12,"license":null,"participation":[0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-22T11:53:35.885Z"}},{"type":"Public","name":"M2PT","owner":"AILab-CVC","isFork":false,"description":"[CVPR'24] Multimodal Pathway: Improve Transformers with Irrelevant Data from Other Modalities","allTopics":["deep-learning","transformers","artificial-intelligence","multimodal"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":86,"forksCount":4,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,2,6,0,0,0,0,0,0,0,2,2,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-13T06:28:04.321Z"}},{"type":"Public","name":"Make-Your-Video","owner":"AILab-CVC","isFork":false,"description":"[IEEE TVCG 2024] Customized Video Generation Using Textual and Structural Guidance","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":2,"starsCount":180,"forksCount":8,"license":"Other","participation":[0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-02-24T07:56:41.852Z"}},{"type":"Public","name":"Animate-A-Story","owner":"AILab-CVC","isFork":false,"description":"Retrieval-Augmented Video Generation for Telling a Story ","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":3,"starsCount":243,"forksCount":17,"license":null,"participation":[0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-02-05T15:07:35.183Z"}},{"type":"Public","name":"UniRepLKNet","owner":"AILab-CVC","isFork":false,"description":"[CVPR'24] UniRepLKNet: A Universal Perception Large-Kernel ConvNet for Audio, Video, Point Cloud, Time-Series and Image Recognition","allTopics":["deep-learning","architecture","artificial-intelligence","convolutional-neural-networks","multimodal-learning"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":5,"starsCount":892,"forksCount":53,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,16,26,3,1,10,5,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-01-02T18:17:46.777Z"}},{"type":"Public","name":"GPT4Tools","owner":"AILab-CVC","isFork":false,"description":"GPT4Tools is an intelligent system that can automatically decide, control, and utilize different visual foundation models, allowing the user to interact with images during a conversation.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":14,"starsCount":752,"forksCount":55,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-12-19T01:55:15.585Z"}},{"type":"Public","name":"GroupMixFormer","owner":"AILab-CVC","isFork":false,"description":"GroupMixAttention and GroupMixFormer","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":3,"starsCount":108,"forksCount":11,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-12-13T05:07:44.494Z"}}],"repositoryCount":17,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"AILab-CVC repositories"}