Skip to content

Commit

Permalink
first commit
Browse files Browse the repository at this point in the history
  • Loading branch information
JosephChenHub committed Dec 10, 2023
1 parent 46f80f9 commit 7f5d501
Show file tree
Hide file tree
Showing 10 changed files with 225 additions and 1 deletion.
Binary file added .DS_Store
Binary file not shown.
1 change: 1 addition & 0 deletions .gitignore
Original file line number Diff line number Diff line change
@@ -0,0 +1 @@
.DS_Store
3 changes: 2 additions & 1 deletion README.md
Original file line number Diff line number Diff line change
@@ -1 +1,2 @@
# gpt4sgg.github.io
# Project Page of GPT4SGG
GPT4SGG: Synthesizing Scene Graphs from Holistic and Region-specific Narratives
102 changes: 102 additions & 0 deletions index.html
Original file line number Diff line number Diff line change
@@ -0,0 +1,102 @@
<!DOCTYPE html>
<html>
<head>
<title>GPT4SGG: Synthesizing Scene Graphs from Holistic and Region-specific Narratives</title>
<link rel="stylesheet" type="text/css" href="style.css">
</head>
<body>
<header>
<div class="header-flex-container">
<img src="resources/polyu-logo.png" alt="PolyU Logo" class="header-logo">
<h1 class="header-title">GPT4SGG: Synthesizing Scene Graphs from Holistic and Region-specific Narratives</h1>
</div>
</header>
<nav>
<div class="author-affiliations">
<span><a href="https://scholar.google.com/citations?user=kuQ_mrAAAAAJ&hl=en">Zuyao Chen</a><sup>1,2</sup></span>,
<span><a href="https://scholar.google.com/citations?user=XujjZmUAAAAJ&hl=en">Jinlin Wu</a><sup>2,3</sup></span>,
<span><a href="https://scholar.google.com/citations?user=cuJ3QG8AAAAJ&hl=en">Zhen Lei</a><sup>2,3</sup></span>,
<span><a href="">Zhaoxiang Zhang</a><sup>2,3</sup></span>,
<span><a href="https://scholar.google.com/citations?user=w2HXPUUAAAAJ&hl=en">Changwen Chen</a><sup>1</sup></span>
<div class="affiliation">
<sup>1</sup> The Hong Kong Polytechnic University
</div>
<a href="https://chenlab.comp.polyu.edu.hk" style="margin:50px;color:#003366;"> https://chenlab.comp.polyu.edu.hk </a>
<div class="affiliation">
<sup>2</sup> Centre for Artificial Intelligence and Robotics, HKISI, CAS
</div>
<div class="affiliation">
<sup>3</sup> NLPR, Institute of Automation, Chinese Academy of Sciences, Beijing, China
</div>
</div>
<div class="buttons">
<a href="https://arxiv.org/pdf/2312.04314.pdf" class="button">arXiv</a>
<a href="" class="button">Code</a>
<a href="" class="button">Demo</a>
<a href="" class="button">Dataset</a>
<a href="" class="button">Model</a>
</div>
</nav>
<section>
<div class="flex-container">
<img src="resources/GPT4SGG-intro.png" style="width:70%;height:auto;" alt="Challenges in learning scene graphs from natural language description.">
</div>
<h2>Abstract</h2>
<p> Learning scene graphs from natural language descriptions has proven to be a cheap and promising scheme for Scene Graph Generation (SGG). However, such unstructured caption data and its processing are troubling the learning an acurrate and complete scene graph. This dilema can be summarized as three points. <ul><li><b>First</b>, traditional language parsers often fail to extract meaningful relationship triplets from caption data.</li><li><b>Second</b>, grounding unlocalized objects in parsed triplets will meet ambiguity in visual-language alignment.</li><li><b>Last</b>, caption data typically are sparse and exhibit bias to partial observations of image content. These three issues make it hard for the model to generate comprehensive and accurate scene graphs.</li></ul> To fill this gap, we propose a simple yet effective framework, <span class="italic-text" style="color: #000080;"><b>GPT4SGG</b></span>, to synthesize scene graphs from holistic and region-specific narratives. The framework discards traditional language parser, and localize objects before obtaining relationship triplets. To obtain relationship triplets, holistic and dense region-specific narratives are generated from the image. With such textual representation of image data and a task-specific prompt, an LLM, particularly GPT-4, directly synthesizes a scene graph as pseudo labels. Experimental results showcase GPT4SGG significantly improves the performance of SGG models trained on image-caption data. We believe this pioneering work can motivate further research into mining the visual reasoning capabilities of LLMs.</p>
</section>
<section>
<h2> Method </h2>
<div class="flex-container">
<img src="resources/GPT4SGG-main.png" style="width:70%; height:auto;" alt="Overview of <b>GPT4SGG</b>.">
</div>
<p><b>Textual representation of image data</b>: localised objects, holistic & region-specific narratives. </p>
<p><b>Task-specific (SGG-aware) Prompt</b>: synthesize scene graphs based on the textual input for image data. </p>
</section>
<section>
<h2> Example of GPT4SGG </h2>
<div class="flex-container">
<img src="resources/GPT4SGG-example.png" style="width:100%; height:auto;" alt="gpt4sgg-example">
</div>
</section>
<section>
<h2> Samples of COCO-SG@GPT </h2>
<div class="flex-container">
<img src="resources/GPT4SGG-samples.png" style="width:100%; height:auto;" alt="gpt4sgg-samples">
</div>
</section>
<section>
<h2> BibTeX </h2>
<p> Please cite <b><span class="italic-text" style="color: #000080;">GPT4SGG</span></b> in your publications if it helps your research: </p>
<button onclick="copyToClipboard()">Copy to Clipboard</button>
<pre><code id="codeBlock">
@misc{chen2023gpt4sgg,
title={GPT4SGG: Synthesizing Scene Graphs from Holistic and Region-specific Narratives},
author={Zuyao Chen and Jinlin Wu and Zhen Lei and Zhaoxiang Zhang and Changwen Chen},
year={2023},
eprint={2312.04314},
archivePrefix={arXiv},
primaryClass={cs.CV}
}</code></pre>
<!-- JavaScript -->
<script>
function copyToClipboard() {
const codeBlock = document.getElementById('codeBlock');
const range = document.createRange();
range.selectNode(codeBlock);
window.getSelection().addRange(range);
try {
document.execCommand('copy');
console.log('Copied to clipboard');
} catch (err) {
console.error('Failed to copy: ', err);
}
window.getSelection().removeAllRanges();
}
</script>
</section>
<footer>
<p>Acknowledgement: ChatGPT for website building</p>
</footer>
</body>
</html>

Binary file added resources/GPT4SGG-example.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Binary file added resources/GPT4SGG-intro.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Binary file added resources/GPT4SGG-main.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Binary file added resources/GPT4SGG-samples.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Binary file added resources/polyu-logo.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
120 changes: 120 additions & 0 deletions style.css
Original file line number Diff line number Diff line change
@@ -0,0 +1,120 @@
body {
font-family: 'Gilmer', sans-serif;
line-height: 1.6;
color: #333;
}


footer {
margin: 20px;
padding: 15px;
border: 1px solid #ddd;
}

section {
margin: 200px;
}

h2 {
text-align: center;
}

header {
display: flex;
align-items: center;
justify-content: space-between; /* Changed to space-between */
}

.header-logo {
width: 250px; /* Adjust as necessary */
height: auto; /* Maintain aspect ratio */
}

.header-title {
font-size: 2em; /* Adjust title size as necessary */
text-align: center;
flex-grow: 1; /* Allows the title to grow and use up available space */
margin: 20px;
margin-left:200px;
padding-left: auto;
padding-right: auto;
}

nav {
color: #003366; /* This is a navy blue */
text-align: center;
}

section {
background-color: #fff;
}

footer {
background-color: #f4f4f4;
text-align: center;
}

.author-affiliations span {
font-weight: bold;
}

.author-affiliations a {
text-decoration: none;
color: black;
}

.affiliation {
margin-top: 5px;
}

.emails {
font-style: italic;
margin-top: 5px;
}


.italic-text {
font-style: italic;
}

.buttons {
margin: 20px;
text-align: center; /* This centers the inline-block elements */
}

.button {
text-decoration: none;
background-color: #4169E1; /* PolyU Red for buttons */
color: white; /* White text */
border: none;
padding: 10px 20px; /* Size of the button */
border-radius: 20px; /* Rounded corners */
font-size: 16px; /* Text size */
cursor: pointer; /* Mouse cursor on hover */
display: inline-block; /* Allows for text-align center to work */
margin: 5px; /* Optional: adds space between the buttons */
}


.button:hover {
background-color: #555; /* Lighter gray on hover */
}


pre {
background-color: #f4f4f4; /* light grey background */
border: 1px solid #ddd; /* light grey border */
padding: 10px;
}

code {
font-family: 'Courier New', Courier, monospace; /* monospaced font */
}


.flex-container {
display: flex;
justify-content: center;
align-items: center;
}

0 comments on commit 7f5d501

Please sign in to comment.