<?xml version='1.0' encoding='utf-8'?>
<?xml-stylesheet type="text/xsl" href="/v2/static/oai2.xsl"?>
<OAI-PMH xmlns="http://www.openarchives.org/OAI/2.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://www.openarchives.org/OAI/2.0/ http://www.openarchives.org/OAI/2.0/OAI-PMH.xsd">
  <responseDate>2026-05-06T03:33:57Z</responseDate>
  <request identifier="oai:figshare.com:article/27004666" metadataPrefix="oai_datacite" verb="GetRecord">https://api.figshare.com/v2/oai</request>
  <GetRecord>
    <record>
      <header>
        <identifier>oai:figshare.com:article/27004666</identifier>
        <datestamp>2024-09-13T14:07:31Z</datestamp>
        <setSpec>category_26275</setSpec>
        <setSpec>category_28840</setSpec>
        <setSpec>category_28849</setSpec>
        <setSpec>category_28879</setSpec>
        <setSpec>category_28885</setSpec>
        <setSpec>category_28888</setSpec>
        <setSpec>portal_549</setSpec>
        <setSpec>item_type_3</setSpec>
        <setSpec>month_year_09_2024</setSpec>
      </header>
      <metadata>
        <resource xmlns="http://datacite.org/schema/kernel-4" xsi:schemaLocation="http://datacite.org/schema/kernel-4 http://schema.datacite.org/meta/kernel-4.3/metadata.xsd">
          <identifier identifierType="DOI">10.5522/04/27004666.v2</identifier>
          <alternateIdentifiers>
            <alternateIdentifier alternateIdentifierType="URL">https://figshare.com/articles/dataset/PitVQA_A_Dataset_of_Visual_Question_Answering_in_Pituitary_Surgery/27004666</alternateIdentifier>
          </alternateIdentifiers>
          <relatedIdentifiers>
            <relatedIdentifier relatedIdentifierType="URL" relationType="HasPart">https://ndownloader.figshare.com/files/49158880</relatedIdentifier>
            <relatedIdentifier relatedIdentifierType="URL" relationType="HasPart">https://ndownloader.figshare.com/files/49228108</relatedIdentifier>
            <relatedIdentifier relatedIdentifierType="URL" relationType="HasPart">https://ndownloader.figshare.com/files/49228111</relatedIdentifier>
            <relatedIdentifier relatedIdentifierType="URL" relationType="IsSupplementedBy">https://github.com/mobarakol/PitVQA</relatedIdentifier>
            <relatedIdentifier relatedIdentifierType="arXiv" relationType="IsSupplementedBy">2409.01184</relatedIdentifier>
            <relatedIdentifier relatedIdentifierType="arXiv" relationType="IsSupplementedBy">2405.13949</relatedIdentifier>
          </relatedIdentifiers>
          <creators>
            <creator>
              <creatorName>Hoque, Mobarak</creatorName>
              <givenName>Mobarak</givenName>
              <familyName>Hoque</familyName>
              <nameIdentifier nameIdentifierScheme="ORCID" schemeURI="http://orcid.org">0000-0002-7162-2822</nameIdentifier>
            </creator>
            <creator>
              <creatorName>Clarkson, Matt</creatorName>
              <givenName>Matt</givenName>
              <familyName>Clarkson</familyName>
            </creator>
            <creator>
              <creatorName>Bano, Sophia</creatorName>
              <givenName>Sophia</givenName>
              <familyName>Bano</familyName>
            </creator>
            <creator>
              <creatorName>Stoyanov, Danail</creatorName>
              <givenName>Danail</givenName>
              <familyName>Stoyanov</familyName>
            </creator>
            <creator>
              <creatorName>Marcus, Hani</creatorName>
              <givenName>Hani</givenName>
              <familyName>Marcus</familyName>
            </creator>
          </creators>
          <titles>
            <title><![CDATA[PitVQA: A Dataset of Visual Question Answering in Pituitary Surgery]]></title>
          </titles>
          <subjects>
            <subject>Biomedical imaging</subject>
            <subject>Intelligent robotics</subject>
            <subject>Natural language processing</subject>
            <subject>Computer vision</subject>
            <subject>Image processing</subject>
            <subject>Multimodal analysis and synthesis</subject>
            <subject>Visual Question Answering (VQA)</subject>
            <subject>large language models in medicine</subject>
            <subject>Large language models (LLMs) in healthcare</subject>
            <subject>Vision Language Models</subject>
            <subject>Pituitary surgery</subject>
            <subject>artificial intelligence analysis</subject>
            <subject>surgical data science</subject>
          </subjects>
          <dates>
            <date dateType="Created">2024-09-18</date>
            <date dateType="Updated">2024-09-18</date>
          </dates>
          <resourceType resourceTypeGeneral="Dataset">Dataset</resourceType>
          <publicationYear>2024</publicationYear>
          <publisher>University College London</publisher>
          <rightsList>
            <rights rightsURI="https://creativecommons.org/licenses/by-nc-nd/4.0/" rightsIdentifier="CC BY-NC-ND 4.0"/>
            <rights rightsURI="http://purl.org/coar/access_right/c_abf2" rightsIdentifier="open access"/>
          </rightsList>
          <descriptions>
            <description descriptionType="Abstract"><![CDATA[<p dir="ltr">PitVQA dataset comprises 25 videos of endoscopic pituitary surgeries from the National Hospital of Neurology and Neurosurgery in London, United Kingdom, similar to the dataset used in the MICCAI PitVis challenge. All patients provided informed consent, and the study was registered with the local governance committee. The surgeries were recorded using a high-definition endoscope (Karl Storz Endoscopy) with a resolution of 720p and stored as MP4 files. All videos were annotated for the surgical phases, steps, instruments present and operation notes guided by a standardised annotation framework, which was derived from a preceding international consensus study on pituitary surgery workflow. Annotation was performed collaboratively by 2 neurosurgical residents with operative pituitary experience and checked by an attending neurosurgeon. We extracted image frames from each video at 1 fps and removed any frames that were blurred or occluded. Ultimately, we obtained a total of 109,173 frames, with the videos of minimum and maximum length yielding 2,443 and 7,179 frames, respectively. We acquired frame-wise question-answer pairs for all the categories of the annotation. Overall, there are 884,242 question-answer pairs from 109,173 frames, which is around 8 pairs for each frame. There are 59 classes overall, including 4 phases, 15 steps, 18 instruments, 3 variations of instruments present in a frame, 5 positions of the instruments, and 14 operation notes in the annotation classes. The length of the questions ranges from a minimum of 7 words to a maximum of 12 words.</p><p dir="ltr">The details description of the original videos can be found at the <a href="https://www.synapse.org/Synapse:syn51232283" target="_blank">MICCAI PitVis challenge </a>and the videos can be directly download from <a href="https://doi.org/10.5522/04/26531686" target="_blank">UCL HDR portal</a>.</p>]]></description>
          </descriptions>
        </resource>
      </metadata>
    </record>
  </GetRecord>
</OAI-PMH>
