Person Re-Identification by Deep Joint Learning of Multi-Loss Classification

introduction

we formulate a method for joint learning of local and global feature selection losses designed to optimise person re-id when using only generic matching metrics such as the L2 distance.即聯(lián)合學(xué)習(xí)局部和全局特征。作者認(rèn)為learning any matching distance metric is intrinsically learn- ing a global feature transformation across domains,所以其實(shí)特征的度量用簡(jiǎn)單的比如L2就可以了,主要應(yīng)該聚焦于特征的提取和表達(dá)。
傳統(tǒng)的手工提取特征主要提取的是局部特征,比如把圖像切分成水平的條狀來(lái)處理。而dl(deep learning)的方法主要提取的是圖像的全局特征。但是作者認(rèn)為這兩種處理方式得到的特征都不是最優(yōu)的,兩者結(jié)合才好,因?yàn)槿说囊曈X(jué)系統(tǒng)是同時(shí)處理這兩種特征(global (contextual) and local (saliency) information)的。仔細(xì)想想,還是有那么點(diǎn)道理的。
作者的網(wǎng)絡(luò)設(shè)計(jì)也是從這個(gè)角度出發(fā),有兩個(gè)branch,分別提取局部特征和全局特征,但是這個(gè)兩個(gè)branch不是獨(dú)立的,而是相互影響,共同學(xué)習(xí)的。這樣一個(gè)網(wǎng)絡(luò)的好處在于,不但可以同時(shí)提取局部和全局的特征,還可以學(xué)習(xí)局部和全局的關(guān)系,兩者相互補(bǔ)足,來(lái)解決局部錯(cuò)位等reID中的典型問(wèn)題。
此外,作者還introduce a structured sparsity based feature selection learning mechanism for improving multi- loss joint feature learning robustness w.r.t. noise and data co- variance between local and global representations.意思大概就是這是一種基于稀疏性的正則化的手段,用來(lái)解決噪聲影響。

related work

1.saliency learning based models。這些方法不考慮全局特征,主要modelling localised part im- portance. However, these existing methods consider only the patch appearance statistics within individual locations but no global feature representation learning, let alone the correla- tion and complementary information discovery between local and global features as modelled by the JLML.
2.Spatially Constrained Similarity (SCS) model和Multi-Channel Parts (MCP) network 。這兩個(gè)方法倒是同時(shí)考慮了全局特征。SCS主要聚焦于 supervised metric learning。但是SCS不考慮hand-crafted local and global features之間的關(guān)系。MCP主要用triplet ranking loss(不懂)來(lái)優(yōu)化,而JLML主要用multiple classification loss,前者存在一定壞處:Critically, this one-loss model learning is likely to impose negative influ- ence on the discriminative feature learning behaviour for both branches due to potential over-low pre-branch independence and over-high inter-branch correlation. This may lead to sub- optimal joint learning of local and global feature selections in model optimisation, as suggested by our evaluation in Section4.3
3.HER model。主要用了regression loss,而JLML主要用的是classification loss。
4.DGD。這篇文章我仔細(xì)看過(guò),它用的也是classification loss。和JLML的區(qū)別在于 他是one-loss classification 而JLML是 multi-loss classifi- cation

模型設(shè)計(jì)

image.png
image.png

(Note that, the ReLU,rectification non-linearity [Krizhevsky et al., 2012] after each conv layer is omitted for brevity.)

兩個(gè)分支分別提取局部和全局特征。聯(lián)合學(xué)習(xí)體現(xiàn)在下面兩個(gè)方面:
1.low level的特征共享。有兩個(gè)好處,第一,共享特征,第二,減少參數(shù),防止過(guò)擬合,尤其是在reID這個(gè)問(wèn)題上,因?yàn)閞eID的數(shù)據(jù)集比較小
2.最后把兩個(gè)512維的特征向量疊加(local and global)

損失函數(shù)

這里他們的損失函數(shù)的選擇不同于大多數(shù)現(xiàn)存的deep reID方法,他們的損失函數(shù)主要用的是 cross- entropy classification loss function。顯存的deep reID方法主要用的contrastive loss,designed to exploit pairwise re-id labels de- fined by both positive and negative pairs, such as the pairwise verification。代表之一是An improved deep learning architecture for person re- identification. In CVPR, 2015.
這么選擇損失函數(shù)的理由如下(不翻譯了,說(shuō)的還挺有道理的):The motivations for our JLML classification loss based learning are: (i) Significantly simplified training data batch construc- tion, e.g. random sampling with no notorious tricks required, as shown by other deep classification methods [Krizhevsky et al., 2012]. This makes our JLML model more scalable in real-world applications with very large training population sizes when available. This also eliminates the undesirable need for carefully forming pairs and/or triplets in preparing re-id training splits, as in most existing methods, due to the inherent imbalanced negative and positive pair size distribu- tions. (ii) Visual psychophysical findings suggest that rep- resentations optimised for classification tasks generalise well to novel categories [Edelman, 1998]. We consider that re- id tasks are about model generalisation to unseen test iden- tity classes given training data on independent seen identity classes. Our JLML model learning exploits this general clas- sification learning principle beyond the strict pair-wise rela- tive verification loss in existing re-id models.大意就是不要用正負(fù)樣本這種形式,直接用正樣本。DGD這篇文章也是用的一樣的思想。

其他

最后就是一些訓(xùn)練細(xì)節(jié),以及對(duì)模型各種方法有和沒(méi)有的比較,證明這些方法是有好處的。好處最明顯的就是聯(lián)合global和local特征了:

image.png

還有就是兩個(gè)分支單獨(dú)學(xué)習(xí)比一起學(xué)習(xí)要好:

image.png

其他的比如有沒(méi)有l(wèi)ow level的shared feature和metric learning的選擇,以及selective feature learning(就是那個(gè)看不懂的正則化),作用甚微。

最后編輯于
?著作權(quán)歸作者所有,轉(zhuǎn)載或內(nèi)容合作請(qǐng)聯(lián)系作者
【社區(qū)內(nèi)容提示】社區(qū)部分內(nèi)容疑似由AI輔助生成,瀏覽時(shí)請(qǐng)結(jié)合常識(shí)與多方信息審慎甄別。
平臺(tái)聲明:文章內(nèi)容(如有圖片或視頻亦包括在內(nèi))由作者上傳并發(fā)布,文章內(nèi)容僅代表作者本人觀點(diǎn),簡(jiǎn)書系信息發(fā)布平臺(tái),僅提供信息存儲(chǔ)服務(wù)。

相關(guān)閱讀更多精彩內(nèi)容

  • 我還是不后悔,愛(ài)你和不愛(ài)你的決定 只是一個(gè)人演的實(shí)在辛苦 我又偏偏生性就怕苦 我還是不后悔,沒(méi)有向你道句珍重 只是...
    珠老三閱讀 335評(píng)論 0 0
  • 最近正打算進(jìn)修,找到一個(gè)網(wǎng)絡(luò)培訓(xùn)班,里面的描述是這樣子的,也許課程開發(fā)者覺(jué)得這樣有說(shuō)服力,說(shuō)服力在哪里呢,當(dāng)然是課...
    浪漫的高貴閱讀 519評(píng)論 6 11
  • 01 周末的這天準(zhǔn)時(shí)地還是在這個(gè)點(diǎn)醒,這已經(jīng)顯然成為一個(gè)習(xí)慣,就像手機(jī)鬧鐘如果你定死了一個(gè)時(shí)間,它就會(huì)在那個(gè)點(diǎn)準(zhǔn)時(shí)...
    黑發(fā)長(zhǎng)衣閱讀 269評(píng)論 0 2
  • 11年來(lái)到這座小城的一個(gè)村子,它挨蒼山很近,如果俯瞰村落,整好是在山坳里,從風(fēng)水學(xué)上看,是個(gè)宜居的地方。晏子還在學(xué)...
    太陽(yáng)上de魚閱讀 265評(píng)論 0 0

友情鏈接更多精彩內(nèi)容