forked from pengzhiliang/pengzhiliang.github.io
-
Notifications
You must be signed in to change notification settings - Fork 1
/
index.html
193 lines (173 loc) · 8.13 KB
/
index.html
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
<!DOCTYPE html>
<HTML>
<HEAD>
<META content="IE=5.0000" http-equiv="X-UA-Compatible">
<META name="description" content="Zonghao Guo's home page">
<META http-equiv="Content-Type" content="text/html; charset=gb2312">
<LINK href="files/doc.css"
rel="stylesheet" type="text/css">
<TITLE>Zonghao Guo</TITLE>
<META name="GENERATOR" content="MSHTML 11.00.10570.1001">
</HEAD>
<BODY>
<DIV id="layout-content" style="margin-top: 25px;">
<TABLE>
<TBODY>
<TR>
<TD width="670">
<DIV id="toptitle">
<H1>Zonghao Guo </H1></DIV>
<H3>Ph.D. candidate</H3>
<P>Room 330, Academy 2 Building
<BR>School of Electronic, Electrical and Communication Engineering
<BR>University of Chinese Academy of Sciences
<BR>Beijing, China, 101408.
<BR>
<BR> Email:
<A href="mailto:[email protected]"> [email protected]</A>;
<BR> Github:
<A href="https://github.com/guozonghao96">https://github.com/guozonghao96</A>;
<BR> Google scholar:
<A href="https://scholar.google.com.hk/citations?hl=&user=h1I6LJcAAAAJ&hl=en">https://scholar.google.com</A>
<BR><BR></P>
</TD>
<TD>
<IMG width="150" src="files/person_photo.jpg" border="0">
</TD>
</TR>
<TR></TR></TBODY>
</TABLE>
<DIV id="layout-content" style="margin-top: 25px;">
<H2>Biography</H2>
<P> I am a Ph.D. candidate of <A href="http://lamp.ucas.ac.cn/">PRISDL</A> in the <A href="http://eece.ucas.ac.cn/index.php/en/">School of Electronic, Electrical and Communication Engineering</A>,
<A href="http://english.ucas.ac.cn/">University of Chinese Academy of Sciences </A>,
advised by <A href="http://people.ucas.ac.cn/~0007279?language=en">Prof. Qixiang Ye</A>.
I got a B.E. degree in Wuhan University, Wuhan in June 2019.
</P>
<P>My research interests include computer vision and deep learning, specifically for object detection and instance segmentation.</P>
<H2>Publications</H2>
<table class="pub_table">
<!-- <tbody> -->
<tr>
<td class="pub_td1"><img src="files/PaperFig/LLaVA-UHD.jpg" class="papericon"></td>
<td
class="pub_td2"><u>Zonghao Guo</u>, Ruyi Xu, Yuan Yao, Junbo Cui, Zanlin Ni, Chunjiang Ge, Tat-Seng Chua, Zhiyuan Liu, Maosong Sun, Gao Huang
<br><b>LLaVA-UHD: an LMMPerceiving Any Aspect Ratio and High-Resolution Images</b>
<br>European Conference on Computer Vision, 2024
<br>
[<a href="https://arxiv.org/pdf/2403.11703">Paper</a>]
[<a href="https://github.com/thunlp/LLaVA-UHD">Code</a>]
<br>
</td>
</tr>
<tr>
<td class="pub_td1"><img src="files/PaperFig/control_cap.png" class="papericon"></td>
<td
class="pub_td2">Yuzhong Zhao, Yue Liu, <u>Zonghao Guo</u>, Weijia Wu, Chen Gong, Fang Wan, Qixiang Ye
<br><b>ControlCap: Controllable Region-level Captioning</b>
<br>European Conference on Computer Vision, 2024
<br>
[<a href="https://arxiv.org/pdf/2401.17910">Paper</a>]
[<a href="https://github.com/callsys/ControlCap">Code</a>]
<br>
</td>
</tr>
<tr>
<td class="pub_td1"><img src="files/PaperFig/cfa.jpg" class="papericon"></td>
<td
class="pub_td2"><u>Zonghao Guo</u>, Chang Liu, Xiaosong Zhang, Jianbin Jiao, Xiangyang Ji, Qixiang Ye
<br><b>Beyond Bounding-Box: Convex-hull Feature Adaptation for Oriented and Densely Packed Object Detection</b>
<br>IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021
<br>
[<a href="https://openaccess.thecvf.com/content/CVPR2021/papers/Guo_Beyond_Bounding-Box_Convex-Hull_Feature_Adaptation_for_Oriented_and_Densely_Packed_CVPR_2021_paper.pdf">Paper</a>]
[<a href="https://github.com/guozonghao96/BeyondBoundingBox">Code</a>]
<br>
</td>
</tr>
<tr>
<td class="pub_td1"><img src="files/PaperFig/Conformer_tpami.png" class="papericon"></td>
<td
class="pub_td2">Zhiliang Peng, <u>Zonghao Guo</u>, Wei Huang, Yaowei Wang, Lingxi Xie, Jianbin Jiao, Qixiang Ye
<br><b>Conformer: Local features coupling global representations for visual recognition and detection</b>
<br>IEEE Transactions on Pattern Analysis and Machine Intelligence, 2023
<br>
[<a href="https://doi.org/10.1109/TPAMI.2023.3243048">Paper</a>]
[<a href="https://github.com/pengzhiliang/Conformer">Code</a>]
<br>
</td>
</tr>
<tr>
<td class="pub_td1"><img src="files/PaperFig/attn_shift.jpg" class="papericon"></td>
<td
class="pub_td2">Mingxiang Liao*, <u>Zonghao Guo*</u>, Yuze Wang, Peng Yuan, Bailan Feng, Fang Wan
<br><b>AttentionShift: Iteratively Estimated Part-based Attention Map for Pointly Supervised Instance Segmentation</b>
<br>IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023
</td>
</tr>
<tr>
<td class="pub_td1"><img src="files/PaperFig/imTED.png" class="papericon"></td>
<td
class="pub_td2">Feng Liu, Xiaosong Zhang, Zhiliang Peng, <u>Zonghao Guo</u>, Fang Wan, Xiangyang Ji, Qixiang Ye
<br><b>Integral Migrating Pre-trained Transformer Encoder-decoders for Visual Object Detection</b>
<br>IEEE International Conference on Computer Vision, 2023
<br>
[<a href="https://arxiv.org/abs/2205.09613">arXiv preprint</a>]
[<a href="https://github.com/LiewFeng/imTED">Code</a>]
</td>
</tr>
<tr>
<td class="pub_td1"><img src="files/PaperFig/cfa_tcsvt.jpg" class="papericon"></td>
<td
class="pub_td2"><u>Zonghao Guo</u>, Xiaosong Zhang, Chang Liu, Xiangyang Ji, Jianbin Jiao, Qixiang Ye
<br><b>Convex-hull Feature Adaptation for Oriented and Densely Packed Object Detection</b>
<br>IEEE Transactions on Circuits and Systems for Video Technology, 2022
<br>
[<a href="https://ieeexplore.ieee.org/document/9668956">Paper</a>]
[<a href="https://github.com/guozonghao96/BeyondBoundingBox">Code</a>]
<br>
</td>
</tr>
<tr>
<td class="pub_td1"><img src="files/PaperFig/bfg.jpg" class="papericon"></td>
<td
class="pub_td2">Yongqiang Mao*, <u>Zonghao Guo*</u>, Xiaonan Lu, Zhiqiang Yuan, Haowen Guo
<br><b>Bidirectional Feature Globalization for Few-shot Semantic Segmentation of 3D Point Cloud Scenes</b>
<br>International Conference on 3D Vision (Oral), 2022
<br>
[<a href="https://arxiv.org/abs/2208.06671">Paper</a>]
<br>
</td>
</tr>
<tr>
<td class="pub_td1"><img src="files/PaperFig/LDA.png" class="papericon"></td>
<td
class="pub_td2">Zhiliang Peng, Wei Huang, <u>Zonghao Guo</u>, Xiaosong Zhang, Jianbin Jiao, Qixiang Ye
<br><b>Long-tailed Distribution Adaptation</b>
<br>29th ACM International Conference on Multimedia, 2021
<br>
[<a href="https://arxiv.org/abs/2110.02686">Paper</a>]
[<a href="https://github.com/pengzhiliang/LDA">Code</a>]
<br>
</td>
</tr>
<tr>
<td class="pub_td1"><img src="files/PaperFig/dgfa.jpg" class="papericon"></td>
<td
class="pub_td2">Yongqiang Mao, Xian Sun, Kaiqiang Chen, Wenhui Diao, <u>Zonghao Guo</u>, Xiaonan Lu, Kun Fu
<br><b>Semantic Segmentation for Point Cloud Scenes via Dilated Graph Feature Aggregation and Pyramid Decoders</b>
<br>AAAI Conference on Artificial Intelligence Workshop, 2021
<br>
[<a href="https://arxiv.org/abs/2204.04944">Paper</a>]
<br>
</td>
</tr>
<!-- </tbody> -->
</table>
<!-- <br>
<H2>Awards</H2>
<LI> Excellent Student Scholarship, Chinese Academy of Sciences, 2020. </LI> -->
<br> <br>
<H2>Statistics</H2>
<script type="text/javascript" src="//rf.revolvermaps.com/0/0/8.js?i=5063gq35g0n&m=0&c=ff0000&cr1=ffffff&f=arial&l=33" async="async"></script>
</BODY>
</HTML>