pl 1
a1140a3d0df1c81e24ae954d935e8926-Supplemental.pdf
PL 1 i=l fRT(xi)) would be, and another term of E XL( Xl PL 1 i=l fRT(Xi)) that propagates through theTransformer blocks. Figure 1 shows the full comparison of the baseline and PLD, fine-tuned at different checkpoints. Overall, we observe that PLD not only trains BERT faster in pre-training but also preserves the performanceondownstreamtasks. Results are visualized in Figure 1, which shows that the baseline is less robust on the choice of learningrates.
)V. (2) MSA is constructed based on Attention by split the channels ofQ,K and V into h groups with each group apart ofqueries, keys,and valuesQi,Ki RN
F s,iBs,i, n = 1,2,...,S, (10) where F s is the support features extracted by a pretrained ViT. Inspired by the multiple-object tracking within a single framework [21], in which different objects are represented by various identifications (i.e., learnable vectors) for simultaneously tracking, we add extra learnable tokens tothemeanfeatures formorediscriminativeprompts.
1fb36c4ccf88f7e67ead155496f02338-Supplemental.pdf
We note that eachxli,t influences Et in two ways: (i) it occurs in Eq.(6) explicitly, but (ii) it also determinesthevaluesof µl 1k,t viaEq.(1). For the experiments on MHNs, the parameterβ was extensively tested, as we have usedβ {1,2,3,5,10,100,1000},and always reported the best result. For experiments comparing against classical Hopfield networks, we have convertedeveryimagetobinary. Results: The results, plotted inFigure 1, are similar tothe ones ofCIFAR10. So far, we analyzed images with Gaussian noise of variance0.2.
Detecting Backdoors in Pre-trained Encoders
Feng, Shiwei, Tao, Guanhong, Cheng, Siyuan, Shen, Guangyu, Xu, Xiangzhe, Liu, Yingqi, Zhang, Kaiyuan, Ma, Shiqing, Zhang, Xiangyu
Self-supervised learning in computer vision trains on unlabeled data, such as images or (image, text) pairs, to obtain an image encoder that learns high-quality embeddings for input data. Emerging backdoor attacks towards encoders expose crucial vulnerabilities of self-supervised learning, since downstream classifiers (even further trained on clean data) may inherit backdoor behaviors from encoders. Existing backdoor detection methods mainly focus on supervised learning settings and cannot handle pre-trained encoders especially when input labels are not available. In this paper, we propose DECREE, the first backdoor detection approach for pre-trained encoders, requiring neither classifier headers nor input labels. We evaluate DECREE on over 400 encoders trojaned under 3 paradigms. We show the effectiveness of our method on image encoders pre-trained on ImageNet and OpenAI's CLIP 400 million image-text pairs. Our method consistently has a high detection accuracy even if we have only limited or no access to the pre-training dataset.
On Kenn's Rule of Combination Applied to Breast Cancer Precision Therapy
Dezert, Jean, Tchamova, Albena
TECHNICAL NOTE - TN-2023-02-28, FEBRUARY 2023. 1 Abstract This short technical note points out an erroneous claim about a new rule of combination of basic belief assignments presented recently by Kenn et al. in [1], referred as Kenn's rule of combination (or just as KRC for short). We prove thanks a very simple counter-example that Kenn's rule is not associative. Consequently, the results of the method proposed by Kenn et al. highly depends on the ad-hoc sequential order chosen for the fusion process as proposed by the authors. This serious problem casts in doubt the interest of this method and its real ability to provide trustful results and to make good decisions to help for precise breast cancer therapy. Recently a paper devoted to the Breast Cancer Precision Therapy by Kenn et al. [1] attracted our attention for two main reasons: 1) this application of information fusion is very interesting and important; 2) Kenn's et al. method is based on a new rule of combination of basic belief assignments (BBAs).
- Health & Medicine > Therapeutic Area > Oncology > Breast Cancer (0.82)
- Health & Medicine > Therapeutic Area > Obstetrics/Gynecology (0.82)