当前位置:网站首页>[point cloud series] unsupervised multi task feature learning on point clouds
[point cloud series] unsupervised multi task feature learning on point clouds
2022-04-23 13:18:00 【^_^ Min Fei】
1. Summary
subject :Unsupervised Multi-Task Feature Learning on Point Clouds
The paper :https://openaccess.thecvf.com/content_ICCV_2019/papers/Hassani_Unsupervised_Multi-Task_Feature_Learning_on_Point_Clouds_ICCV_2019_paper.pdf
2. motivation
Traditional manual design features mainly capture local or global statistical attributes , Unable to represent semantic properties .
Is there a common feature for tasks that require semantic features ?
So this article This paper puts forward a method which can be used for unsupervised multitasking Feature expression of the model , Encoder based on multiscale graph . End to end training .
3. Algorithm
For three unsupervised learning tasks :
- clustering
- restructure
- Self supervised classification
The overall framework :
decoder : Multi scale ( Graph convolution + Convolution + Pooling )+ Perturbed Gaussian noise
Essentially applicable to multitasking , It's also because we used the loss of these tasks to train together , Naturally, with this function .
Definition of volume of drawing : It's actually the residual between two points ;
Algorithm description :
Training set : S = { s 1 , s 2 , . . . , s N } S=\{s_1, s_2, ..., s_N\} S={
s1,s2,...,sN},N A little bit .
One point : s i = { p 1 i , p 2 i , . . , p M 1 } s_i =\{ p^i_1, p^i_2, .., p^1_M\} si={
p1i,p2i,..,pM1}, M A disordered point , p j i = ( x j i , y j i , z j i ) p^i_j=(x^i_j, y^i_j, z^i_j) pji=(xji,yji,zji) Include coordinates only
Encoder : E θ : S ( R M × d i n ) → Z ( R d z ) E_{\theta}: S (\mathbb{R}^{M\times d_{in}})\rightarrow Z (\mathbb{R}^{d_z}) Eθ:S(RM×din)→Z(Rdz), d z d_z dz Far greater than d i n d_in din
In order to learn more tasks than supervision θ \theta θ, Design the following three parameter functions :
Clustering function T c : Z → y \Tau_c:Z \rightarrow y Tc:Z→y, Classify hidden codes into K K K Of the three categories , among y = [ y 1 , y 2 , . . . , y n ] y=[y_1, y_2, ...,y_n] y=[y1,y2,...,yn], y i ∈ { 0 , 1 } K y_i\in\{0,1\}^K yi∈{
0,1}K, And y n T 1 k = 1 y^T_n \mathbf{1}_k=1 ynT1k=1.
Classification function f ψ : Z → y ^ f_\psi: Z \rightarrow \hat{y} fψ:Z→y^, Category prediction after clustering , In other words , The classification function maps implicit variables to K A prediction class y ^ = [ y 1 ^ , y 2 ^ , . . . , y n ^ ] \hat{y}=[\hat{y_1}, \hat{y_2}, ...,\hat{y_n}] y^=[y1^,y2^,...,yn^], And y i ^ ∈ { 0 , 1 } K \hat{y_i}\in\{0,1\}^K yi^∈{
0,1}K. The function uses the pseudo tag generated by the clustering function as the agent training data .
Decoder function : g ϕ : Z ( R d z ) → S ^ ( R M × d i n ) g_\phi: Z(\mathbb{R}^{d_z}) \rightarrow \hat{S} (\mathbb{R}^{M\times d_{in}}) gϕ:Z(Rdz)→S^(RM×din), Reconstruct the implicit variable back to the point cloud . If only clustering loss is used, the features will be clustered into a single class , The function is designed to prevent the final aggregation into a single class .
Loss of training :
-
Clustering loss : In essence, it is learning the clustering center matrix : C ∈ R d z × K C\in \mathbb{R}^{d_z \times K} C∈Rdz×K, z n = E θ ( s n ) z_n=E_\theta(sn) zn=Eθ(sn), y n T 1 k = 1 y^T_n\mathbf{1}_k=1 ynT1k=1. among , initialization The random clustering matrix is the center , yes epoch The updated .
-
Classified loss : Cross entropy measures , y n = T c ( z n ) y_n=\Tau_c(z_n) yn=Tc(zn), y ^ n = f ψ ( z n ) \hat{y}_n=f_\psi(z_n) y^n=fψ(zn).
-
Refactoring loss :CD distance . s ^ n = g ϕ ( z n ) \hat{s}_n=g_\phi (z_n) s^n=gϕ(zn) Is the reconstructed point set , s n s_n sn yes GT. N N N: Number of training sets ; M M M: The number of points in each point set .
The ultimate loss : above 3 Weighted summation of losses
The specific description process is shown in the following figure 1 Shown :
4. experiment
Model convergence :88 Classes , actual 55 Category .
The Clustering Visualization diagram is as follows :
On classified tasks , Comparison of unsupervised and supervised methods : It works well in unsupervised methods ;
The effect of semi supervised segmentation task : 5% The effect of the training set is still very good ;
Split task effect :
Encoder The role of , Test on classification tasks : In fact, it seems to be saying here , complex Encoder The improvement after refactoring is not particularly great , The improvement of multi class tasks is very big .
Ablation Experiment : Refactoring is the key , Plus the effect of classification > The effect of clustering , Because clustering here is actually for further classification , Realize the so-called unsupervised .
5. Conclusion and thinking
Failure case list :
- apply K-Means To initialize the cluster center , But there is no improvement in the random clustering center method ;
- stay decoder And classification model soft Shared parameter mechanism , It is found that the effect is reduced , So separate the two ;
- Try to iterate over more layers , Recalculate the nearest neighbor at each layer , It is found that this is disadvantageous to classification and segmentation ;
版权声明
本文为[^_^ Min Fei]所创,转载请带上原文链接,感谢
https://yzsam.com/2022/04/202204230611136498.html
边栏推荐
- Uninstall MySQL database
- 5道刁钻的Activity生命周期面试题,学完去吊打面试官!
- Xi'an CSDN signed a contract with Xi'an Siyuan University, opening a new chapter in IT talent training
- 5 tricky activity life cycle interview questions. After learning, go and hang the interviewer!
- Armv8m (cortex M33) MPU actual combat
- [51 single chip microcomputer traffic light simulation]
- Async void provoque l'écrasement du programme
- Melt reshape decast long data short data length conversion data cleaning row column conversion
- mui picker和下拉刷新冲突问题
- "Xiangjian" Technology Salon | programmer & CSDN's advanced road
猜你喜欢
【官宣】长沙软件人才实训基地成立!
"Xiangjian" Technology Salon | programmer & CSDN's advanced road
100 lectures on practical application cases of Excel (VIII) - report connection function of Excel
three. JS text ambiguity problem
Riscv MMU overview
AUTOSAR from introduction to mastery lecture 100 (84) - Summary of UDS time parameters
Solve the problem that Oracle needs to set IP every time in the virtual machine
Request和Response及其ServletContext总结
[official announcement] Changsha software talent training base was established!
Ding ~ your scholarship has arrived! C certified enterprise scholarship list released
随机推荐
Mui + hbuilder + h5api simulate pop-up payment style
51 single chip microcomputer stepping motor control system based on LabVIEW upper computer (upper computer code + lower computer source code + ad schematic + 51 complete development environment)
three. JS text ambiguity problem
uniapp image 引入本地图片不显示
How do ordinary college students get offers from big factories? Ao Bing teaches you one move to win!
5道刁钻的Activity生命周期面试题,学完去吊打面试官!
XML
Brief introduction of asynchronous encapsulation interface request based on uniapp
Learning notes of AMBA protocol
100 lectures on practical application cases of Excel (VIII) - report connection function of Excel
8086 of x86 architecture
Design of body fat detection system based on 51 single chip microcomputer (51 + OLED + hx711 + US100)
STD:: shared of smart pointer_ ptr、std::unique_ ptr
The difference between string and character array in C language
AUTOSAR from introduction to mastery 100 lectures (83) - bootloader self refresh
Complete project data of UAV apriltag dynamic tracking landing based on openmv (LabVIEW + openmv + apriltag + punctual atom four axes)
PyTorch 21. NN in pytorch Embedding module
在 pytorch 中加载和使用图像分类数据集 Fashion-MNIST
100 GIS practical application cases (53) - making three-dimensional image map as the base map of urban spatial pattern analysis
The project file '' has been renamed or is no longer in the solution, and the source control provider associated with the solution could not be found - two engineering problems