[1] SHEN D, WU G, SUK H I. Deep learning in medical image analysis[J]. Annual Review of Biomedical Engineering, 2017, 19: 221-248.
[2] CARUANA R. Multitask learning[J]. Machine Learning, 1997, 28(1): 41-75.
[3] GONG T, LEE T, STEPHENSON C, et al. A comparison of loss weighting strategies for multitask learning in deep neural networks[J]. IEEE Access, 2019, 7: 141627-141632.
[4] KENDALL A, GAL Y, CIPOLLA R. Multi-task learning using uncertainty to weigh losses for scene geometry and semantics[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2018: 7482-7491.
[5] CHEN Z, BADRINARAYANAN V, LEE C Y, et al. Gradnorm: Gradient normalization for adaptive loss balancing in deep multitask networks[C]//International Conference on Machine Learning. PMLR, 2018: 794-803.
[6] SENER O, KOLTUN V. Multi-task learning as multi-objective optimization[J]. Advances in Neural Information Processing Systems, 2018, 31.
[7] GUO M, HAQUE A, HUANG D A, et al. Dynamic task prioritization for multitask learning[C]//Proceedings of the European Conference on Computer Vision (ECCV). 2018: 270-287.
[8] YU T, KUMAR S, GUPTA A, et al. Gradient surgery for multi-task learning[J]. Advances in Neural Information Processing Systems, 2020, 33: 5824-5836.
[9] ANDO R K, ZHANG T, BARTLETT P. A framework for learning predictive structures from multiple tasks and unlabeled data[J]. Journal of Machine Learning Research, 2005, 6(11).
[10] ARGYRIOU A, EVGENIOU T, PONTIL M. Multi-task feature learning[J]. Advances in Neural Information Processing Systems, 2006, 19.
[11] OBOZINSKI G, TASKAR B, JORDAN M. Multi-task feature selection[R]. Department of Statistics, University of California, Berkeley, 2006.
[12] JACOB L, VERT J P, BACH F. Clustered multi-task learning: A convex formulation[J]. Advances in Neural Information Processing Systems, 2008, 21.
[13] BAKKER B, HESKES T. Task clustering and gating for bayesian multitask learning[J]. Journal of Machine Learning Research, 2003, 4: 83-99.
[14] BONILLA E V, CHAI K, WILLIAMS C. Multi-task gaussian process prediction[J]. Advances in Neural Information Processing Systems, 2007, 20.
[15] XUE Y, LIAO X, CARIN L, et al. Multi-task learning for classification with dirichlet process priors[J]. Journal of Machine Learning Research, 2007, 8(1).
[16] ZHANG Y, YEUNG D Y, XU Q. Probabilistic multi-task feature selection[J]. Advances in Neural Information Processing Systems, 2010, 23.
[17] MISRA I, SHRIVASTAVA A, GUPTA A, et al. Cross-stitch networks for multi-task learning[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2016: 3994-4003.
[18] LIU P, QIU X, HUANG X J. Adversarial multi-task learning for text classification[C]//Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics. 2017: 1-10.
[19] LONG M, CAO Z, WANG J, et al. Learning multiple tasks with multilinear relationship networks[J]. Advances in Neural Information Processing Systems, 2017, 30.
[20] YANG Y, HOSPEDALES T. Deep multi-task representation learning: A tensor factorization approach[C]//5th International Conference on Learning Representations. 2017.
[21] BAXTER J. A bayesian/information theoretic model of learning to learn via multiple task sampling[J]. Machine Learning, 1997, 28(1): 7-39.
[22] DUONG L, COHN T, BIRD S, et al. Low resource dependency parsing: Cross-lingual parameter sharing in a neural network parser[C]//Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing. 2015: 845-850.
[23] YANG Y, HOSPEDALES T. Trace norm regularised deep multi-task learning[C]//5th International Conference on Learning Representations. 2017.
[24] LECUN Y, BOTTOU L, BENGIO Y, et al. Gradient-based learning applied to document recognition[J]. Proceedings of the IEEE, 1998, 86(11): 2278-2324.
[25] KRIZHEVSKY A, SUTSKEVER I, HINTON G E. Imagenet classification with deep convolutional neural networks[J]. Advances in Neural Information Processing Systems, 2012, 25.
[26] HE K, ZHANG X, REN S, et al. Deep residual learning for image recognition[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2016: 770-778.
[27] HOWARD A G, ZHU M, CHEN B, et al. Mobilenets: Efficient convolutional neural networks for mobile vision applications[A]. 2017.
[28] LIU S, JOHNS E, DAVISON A J. End-to-end multi-task learning with attention[C]//Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2019: 1871-1880.
[29] CHENNUPATI S, SISTU G, YOGAMANI S, et al. Multinet++: Multi-stream feature aggregation and geometric loss strategy for multi-task learning[C]//Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops. 2019: 0-0.
[30] MEHTA N, LEE D, GRAY A. Minimax multi-task learning and a generalized loss compositional paradigm for mtl[J]. Advances in Neural Information Processing Systems, 2012, 25.
[31] BENGIO Y, LOURADOUR J, COLLOBERT R, et al. Curriculum learning[C]//Proceedings of the 26th Annual International Conference on Machine Learning. 2009: 41-48.
[32] KUMAR M, PACKER B, KOLLER D. Self-paced learning for latent variable models[J]. Advances in Neural Information Processing Systems, 2010, 23.
[33] PENTINA A, SHARMANSKA V, LAMPERT C H. Curriculum learning of multiple task[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2015: 5492-5500.
[34] LI C, YAN J, WEI F, et al. Self-paced multi-task learning[C]//Thirty-First AAAI Conference on Artificial Intelligence. 2017.
[35] MURUGESAN K, CARBONELL A. Self-paced multitask learning with shared knowledge[C]//Proceedings of the 26th International Joint Conference on Artificial Intelligence. 2017: 2522-2528.
[36] LIN T Y, GOYAL P, GIRSHICK R, et al. Focal loss for dense object detection[C]//Proceedings of the IEEE International Conference on Computer Vision. 2017: 2980-2988.
[37] LIN X, ZHEN H, LI Z, et al. Pareto multi-task learning[C]//Advances in Neural Information Processing Systems. 2019: 12037-12047.
[38] MAHAPATRA D, RAJAN V. Multi-task learning with user preferences: Gradient descent with controlled ascent in pareto optimization[C]//International Conference on Machine Learning. PMLR, 2020: 6597-6607.
[39] DÉSIDÉRI J A. Multiple-gradient descent algorithm (mgda) for multiobjective optimization[J]. Comptes Rendus Mathematique, 2012, 350(5-6): 313-318.
[40] ZHANG Z, LUO P, LOY C C, et al. Facial landmark detection by deep multi-task learning[C]//European Conference on Computer Vision. Springer, 2014: 94-108.
[41] GAO Y, MA J, ZHAO M, et al. Nddr-cnn: Layerwise feature fusing in multi-task cnns by neural discriminative dimensionality reduction[C]//Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2019: 3205-3214.
[42] ROSENBAUM C, KLINGER T, RIEMER M. Routing networks: Adaptive selection of non-linear functions for multi-task learning[C]//International Conference on Learning Representations. 2018.
[43] FRANCESCHI L, FRASCONI P, SALZO S, et al. Bilevel programming for hyperparameter optimization and meta-learning[C]//International Conference on Machine Learning. PMLR, 2018: 1568-1577.
[44] SAENKO K, KULIS B, FRITZ M, et al. Adapting visual category models to new domains[C]//European Conference on Computer Vision. Springer, 2010: 213-226.
[45] VENKATESWARA H, EUSEBIO J, CHAKRABORTY S, et al. Deep hashing network for unsupervised domain adaptation[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2017: 5018-5027.
[46] RUSSAKOVSKY O, DENG J, SU H, et al. Imagenet large scale visual recognition challenge[J]. International Journal of Computer Vision, 2015, 115(3): 211-252.
[47] KINGMA D P, BA J. Adam: A method for stochastic optimization[C]//3rd International Conference on Learning Representations. 2015.
[48] CORDTS M, OMRAN M, RAMOS S, et al. The cityscapes dataset for semantic urban scene understanding[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2016: 3213-3223.
[49] SILBERMAN N, HOIEM D, KOHLI P, et al. Indoor segmentation and support inference from rgbd images[C]//European Conference on Computer Vision. Springer, 2012: 746-760.
[50] ZHANG Y, YANG Q. A survey on multi-task learning[J]. IEEE Transactions on Knowledge and Data Engineering, 2021.
[51] SANDLER M, HOWARD A, ZHU M, et al. Mobilenetv2: Inverted residuals and linear bottlenecks[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2018: 4510-4520.
[52] HINTON G, DENG L, YU D, et al. Deep neural networks for acoustic modeling in speech recognition: The shared views of four research groups[J]. IEEE Signal Processing Magazine, 2012, 29(6): 82-97.
[53] MCMAHAN B, MOORE E, RAMAGE D, et al. Communication-efficient learning of deep networks from decentralized data[C]//Artificial Intelligence and Statistics. PMLR, 2017: 1273-1282.
[54] LI T, SAHU A K, ZAHEER M, et al. Federated optimization in heterogeneous networks[C]//Proceedings of Machine Learning and Systems. 2020.
[55] KARIMIREDDY S P, KALE S, MOHRI M, et al. Scaffold: Stochastic controlled averaging for federated learning[C]//International Conference on Machine Learning. PMLR, 2020: 5132-5143.
[56] ACAR D A E, ZHAO Y, MATAS R, et al. Federated learning based on dynamic regularization[C]//International Conference on Learning Representations. 2020.
[57] FINN C, ABBEEL P, LEVINE S. Model-agnostic meta-learning for fast adaptation of deep networks[C]//International Conference on Machine Learning. PMLR, 2017: 1126-1135.
[58] FALLAH A, MOKHTARI A, OZDAGLAR A. Personalized federated learning with theoretical guarantees: A model-agnostic meta-learning approach[J]. Advances in Neural Information Processing Systems, 2020, 33: 3557-3568.
[59] COLLINS L, HASSANI H, MOKHTARI A, et al. Exploiting shared representations for personalized federated learning[C]//International Conference on Machine Learning. PMLR, 2021: 2089-2099.
[60] LI X, JIANG M, ZHANG X, et al. Fedbn: Federated learning on non-iid features via local batch normalization[C]//International Conference on Learning Representations. 2020.
[61] LI T, HU S, BEIRAMI A, et al. Ditto: Fair and robust federated learning through personalization[C]//International Conference on Machine Learning. PMLR, 2021: 6357-6368.
[62] SMITH V, CHIANG C K, SANJABI M, et al. Federated multi-task learning[J]. Advances in Neural Information Processing Systems, 2017, 30.
[63] GUO M H, XU T X, LIU J J, et al. Attention mechanisms in computer vision: A survey[J]. Computational Visual Media, 2022: 1-38.
[64] HU J, SHEN L, SUN G. Squeeze-and-excitation networks[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2018: 7132-7141.
[65] WOO S, PARK J, LEE J Y, et al. Cbam: Convolutional block attention module[C]//Proceedings of the European Conference on Computer Vision (ECCV). 2018: 3-19.
[66] WANG J, CHEN Y, CHAKRABORTY R, et al. Orthogonal convolutional neural networks[C]//Proceedings of the IEEE/CVF Conference on Cmputer Vision and Pattern Recognition. 2020: 11505-11515.
[67] HOU Q, ZHOU D, FENG J. Coordinate attention for efficient mobile network design[C]//Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2021: 13713-13722.
[68] WANG X, LI L, YE W, et al. Transferable attention for domain adaptation[C]//Proceedings of the AAAI Conference on Artificial Intelligence: volume 33. 2019: 5345-5352.
[69] WANG Y, ZHANG Z, HAO W, et al. Attention guided multiple source and target domain adaptation[J]. IEEE Transactions on Image Processing, 2020, 30: 892-906.
[70] NETZER Y, WANG T, COATES A, et al. Reading digits in natural images with unsupervised feature learning[C]//NIPS Workshop on Deep Learning and Unsupervised Feature Learning. 2011.
[71] HULL J J. A database for handwritten text recognition research[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 1994, 16(5): 550-554.
[72] GANIN Y, LEMPITSKY V. Unsupervised domain adaptation by backpropagation[C]//International Conference on Machine Learning. PMLR, 2015: 1180-1189.
[73] PENG X, BAI Q, XIA X, et al. Moment matching for multi-source domain adaptation[C]//Proceedings of the IEEE/CVF International Conference on Computer Vision. 2019: 1406-1415.
[74] EVERINGHAM M, WINN J. The pascal visual object classes challenge 2012 (voc2012) development kit[J]. Pattern Analysis, Statistical Modelling and Computational Learning, Tech. Rep, 2011, 8: 5.
[75] DAQUAN Z, HOU Q, CHEN Y, et al. Rethinking bottleneck structure for efficient mobile network design[M]//European Conference on Computer Vision. 2020.
[76] BOYD S, BOYD S P, VANDENBERGHE L. Convex optimization[M]. Cambridge University Press, 2004.
[77] MCDIARMID C, et al. On the method of bounded differences[J]. Surveys in Combinatorics, 1989, 141(1): 148-188.
修改评论