René's URL Explorer Experiment


Title: LLaVA-NeXT: Improved reasoning, OCR, and world knowledge | LLaVA

Open Graph Title: LLaVA-NeXT: A Strong Zero-shot Video Understanding Model

X Title: LLaVA-NeXT: A Strong Zero-shot Video Understanding Model

Description: LLaVA team presents LLaVA-NeXT, with improved reasoning, OCR, and world knowledge. LLaVA-NeXT even exceeds Gemini Pro on several benchmarks.

Opengraph URL: https://llava-vl.github.io/blog/2024-04-30-llava-next-video/

Generator: Jekyll v3.9.4

direct link

Domain: llava-vl.github.io


Hey, it has json ld scripts:
        {
            "@context": "https://schema.org",
            "@type": "BlogPosting",
            "author": {
                "@type": "Person",
                "name": "Yuanhan Zhang, Bo Li, Haotian Liu, Yong Jae Lee, Liangke Gui, Di Fu, Jiashi Feng, Ziwei Liu, Chunyuan Li"
            },
            "dateModified": "2024-01-30T12:33:38-06:00",
            "datePublished": "2024-01-30T12:33:38-06:00",
            "description": "LLaVA team presents LLaVA-NeXT, with improved reasoning, OCR, and world knowledge. LLaVA-NeXT even exceeds Gemini Pro on several benchmarks.",
            "headline": "LLaVA-NeXT: Improved reasoning, OCR, and world knowledge",
            "mainEntityOfPage": {
                "@type": "WebPage",
                "@id": "https://llava-vl.github.io/blog/2024-01-30-llava-next/"
            },
            "url": "https://llava-vl.github.io/blog/2024-01-30-llava-next/"
        }
    

NoneIE=edge
authorYuanhan Zhang, Bo Li, Haotian Liu, Yong Jae Lee, Liangke Gui, Di Fu, Jiashi Feng, Ziwei Liu, Chunyuan Li
og:localeen_US
og:site_nameLLaVA
og:typearticle
article:published_time2024-04-30T12:33:38-06:00
twitter:cardsummary

Links:

LLaVAhttps://llava-vl.github.io/blog/
LLaVA-NeXThttps://llava-vl.github.io/blog/2024-01-30-llava-next/
SGLanghttps://github.com/sgl-project/sglang
Codehttps://github.com/LLaVA-VL/LLaVA-NeXT-Video
Modelhttps://huggingface.co/collections/lmms-lab/llava-next-video-661e86f5e8dabc3ff793c944
Gemini Ultrahttps://blog.google/technology/ai/google-gemini-next-generation-model-february-2024/
VideoChatGPThttps://arxiv.org/abs/2306.05424
Video-LLaVAhttps://arxiv.org/abs/2311.10122
VISTA-LLAMAhttps://arxiv.org/abs/2312.08870
VideoChat2 https://arxiv.org/pdf/2311.17005v3.pdf
LLaMA-VIDhttps://github.com/dvlab-research/LLaMA-VID
LLaMA-VIDhttps://github.com/dvlab-research/LLaMA-VID
our recent study LLaVA-Houndhttps://arxiv.org/abs/2404.01258
Yuanhan Zhanghttps://zhangyuanhan-ai.github.io/
Bo Lihttps://brianboli.com/
Haotian Liuhttps://hliu.cc/
Yong Jae Leehttps://pages.cs.wisc.edu/~yongjaelee/
Liangke Guihttps://scholar.google.com/citations?user=oM6Pj3MAAAAJ&hl=en
Jiashi Fenghttps://sites.google.com/site/jshfeng/home
Ziwei Liuhttps://liuziwei7.github.io/
Chunyuan Lihttps://chunyuan.li/
LLaVA-NeXT: Stronger LLMs Supercharge Multimodal Capabilities in the Wildhttps://llava-vl.github.io/blog/2024-05-10-llava-next-stronger-llms/
LLaVA-NeXT: Improved reasoning, OCR, and world knowledgehttps://llava-vl.github.io/blog/2024-01-30-llava-next/
Accelerating the Development of Large Multimodal Models with LMMs-Evalhttps://lmms-lab.github.io/lmms-eval-blog/lmms-eval-0.1/
https://llava-vl.github.io/blog/2024-01-30-llava-next/

Viewport: width=device-width, initial-scale=1


URLs of crawlers that visited me.