MarkTechPost@AI 2024年07月23日
HuggingFace Researchers Introduce Docmatix: A Dataset For Document Visual Question Answering Containing 2.4 Million Pictures And 9.5 Million Q/A Pairs
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

HuggingFace 研究人员发布了 Docmatix,一个包含 240 万张图片和 950 万个问答对的文档视觉问答数据集,旨在解决 DocVQA 数据集稀缺的问题。Docmatix 从 130 万份 PDF 文档中提取数据,并经过严格的过滤和质量控制,确保数据集的可靠性和高质量。该数据集的发布将推动 DocVQA 模型的开发和应用,并促进文档相关任务的自动化和更广泛的访问。

🤔 Docmatix 是一个包含 240 万张图片和 950 万个问答对的文档视觉问答数据集,由 HuggingFace 研究人员创建,旨在解决 DocVQA 数据集稀缺的问题。

📚 Docmatix 从 130 万份 PDF 文档中提取数据,并经过严格的过滤和质量控制,确保数据集的可靠性和高质量。研究人员使用 Phi-3-small 模型生成问答对,并通过去除包含“不可回答”等关键词的答案来确保数据集的质量。

📈 Docmatix 的发布将推动 DocVQA 模型的开发和应用,并促进文档相关任务的自动化和更广泛的访问。研究人员通过对 Florence-2 模型进行消融实验,证明了 Docmatix 在提高 DocVQA 模型性能方面的潜力。

🚀 Docmatix 的发布将为研究人员和开发者提供一个宝贵的资源,促进 DocVQA 领域的发展。该数据集的规模和质量将推动更先进的 DocVQA 模型的开发,并为文档相关任务的自动化和更广泛的访问提供新的可能性。

🤝 该数据集的发布也体现了 HuggingFace 在推动人工智能领域开放研究和合作方面的努力。通过提供高质量的数据集,HuggingFace 促进了人工智能技术的发展和应用,并为更广泛的社区提供了更多机会。

💻 Docmatix 数据集已在 Hugging Face Hub 上公开发布,供研究人员和开发者使用。

🔓 该数据集的发布将有助于解决 DocVQA 领域面临的挑战,并推动该领域的发展。

🔍 Docmatix 数据集的发布将促进对文档相关任务的更深入的研究,并为更广泛的应用场景提供新的可能性。

💡 Docmatix 数据集的发布将为人工智能技术在文档处理和信息提取方面的应用提供新的动力。

🧠 Docmatix 数据集的发布将推动人工智能技术在文档理解和知识获取方面的进步。

🌐 Docmatix 数据集的发布将促进人工智能技术在跨领域和跨语言的文档处理方面的应用。

🚀 Docmatix 数据集的发布将为人工智能技术在文档自动化的发展开辟新的道路。

🎉 Docmatix 数据集的发布将为人工智能技术在文档管理和信息检索方面的应用带来新的机遇。

🚀 Docmatix 数据集的发布将推动人工智能技术在文档分析和信息挖掘方面的进步。

💡 Docmatix 数据集的发布将为人工智能技术在文档理解和知识推理方面的应用提供新的可能。

🚀 Docmatix 数据集的发布将推动人工智能技术在文档自动化和智能化方面的发展。

🎉 Docmatix 数据集的发布将为人工智能技术在文档处理和信息服务方面的应用带来新的机遇。

🚀 Docmatix 数据集的发布将推动人工智能技术在文档分析和信息提取方面的进步。

💡 Docmatix 数据集的发布将为人工智能技术在文档理解和知识获取方面的应用提供新的动力。

🌐 Docmatix 数据集的发布将促进人工智能技术在跨领域和跨语言的文档处理方面的应用。

🚀 Docmatix 数据集的发布将为人工智能技术在文档自动化的发展开辟新的道路。

🎉 Docmatix 数据集的发布将为人工智能技术在文档管理和信息检索方面的应用带来新的机遇。

Document Visual Question Answering (DocVQA) is a branch of visual question answering that focuses on answering queries about the contents of documents. These documents can take several forms, including scanned photographs, PDFs, and digital documents with text and visual features. However, there are few datasets for DocVQA because collecting and annotating the data is complicated. It requires understanding the context, structure, and layout of various document formats, which requires much manual effort. Due to the sensitive nature of the information contained within, many documents are inaccessible or have privacy concerns that make sharing or using them difficult. Domain-specific differences and the absence of document-structure uniformity further complicate the development of an exhaustive dataset. Factors contributing to the complexity of multi-modal fusion and the accuracy of optical character recognition also play a role. 

Despite these challenges, the urgent need for more DocVQA datasets is underscored. These datasets are crucial for enhancing model performance, as they enable more thorough benchmarking and enhance model training for higher generalizability. By automating document-related processes across sectors and making documents more accessible through summary generation and query responding, updated DocVQA models could significantly impact document accessibility.

To fine-tune Vision-Language Models (VLMs), and Idefics2 in particular, researchers from HuggingFace initially built The Cauldron, a massive collection of fifty datasets. As a result of these efforts, the team discovered a severe shortage of high-quality datasets for Document Visual Question Answering (DocVQA). With 10,000 photos and 39,000 question-answer (Q/A) pairings, DocVQA was the main dataset used for Idefics2. There still needs to be a significant performance disparity between open-source and closed-source models, even after fine-tuning this and other datasets.

Their new study introduces Docmatix, a monumental DocVQA dataset containing 2.4 million pictures and 9.5 million Q/A pairs extracted from 1.3 million PDF documents. This scale, which has increased by 240 times compared to earlier datasets, showcases the potential impact of Docmatix.

The PDFA collection, which includes over two million PDFs, is the source of Docmatix. The researchers used a Phi-3-small model to create Q/A pairs using the PDFA transcriptions. To make sure the dataset was good, 15% of the Q/A pairings were removed that were found to be hallucinations during the creation filter. This was accomplished by eliminating responses that included the word “unanswerable” using regular expressions that detect code. There is a row in the dataset for every PDF. After processing the PDFs, the team saved 150 dpi photographs to the Hugging Face Hub. Now, anyone may access them with ease. 

Users can place their full trust in Docmatix, as all PDFs can be traced back to the original PDFA dataset. Despite the resource-intensive process of converting several PDFs to photos, the researchers have uploaded the processed images for user convenience.

After processing the initial small dataset batch, the researchers ran multiple ablation experiments to fine-tune the prompts. They were aiming for approximately four question-and-answer pairs per page. A few pairs lack detail, whereas excess pairs indicate high overlap. Furthermore, they strived for responses resembling human speech, meaning they were neither lengthy nor brief. To avoid duplicating efforts, the questions were diverse. Surprisingly, there were few instances of question repetition when the Phi-3 model was instructed to inquire about certain details in the text (for example, “What are the titles of John Doe?”).

The team used the Florence-2 model to undertake ablation trials to assess Docmatix’s performance. To facilitate comparability, they trained a pair of versions of the model. The DocVQA dataset was used to train the initial version over multiple epochs. To ensure the model produced the right format for DocVQA evaluation, the second version was trained for one epoch on Docmatix (20% of the images and 4% of the Q/A pairs) and then for one epoch on DocVQA. The findings are noteworthy: a relative improvement of about 20% was produced by training on this tiny subset of Docmatix. Although much bigger, the 0.7B Florence-2 model only did 5% worse than the 8B Idefics2 model trained on various datasets.

The team hopes their work reduces the disparity between proprietary and open-sourced VLMs. To train a brand new, fantastic DocVQA model, they urge the open-source community to use Docmatix. 


Check out the Dataset and Details. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. If you like our work, you will love our newsletter..

Don’t Forget to join our 46k+ ML SubReddit

Find Upcoming AI Webinars here

The post HuggingFace Researchers Introduce Docmatix: A Dataset For Document Visual Question Answering Containing 2.4 Million Pictures And 9.5 Million Q/A Pairs appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

Docmatix 文档视觉问答 DocVQA 数据集 HuggingFace 人工智能 深度学习 自然语言处理 计算机视觉
相关文章