Smart Fashion is reshaping people's lives, and affects people's choices and outfits. Existing computer-vision-enabled fashion technology has covered many aspects, such as fashion detection, fashion recognition, fashion segmentation, virtual fitting, fashion recommendation and fashion compatibility, etc. However, there is a gap in the direction of intelligent care. The care process is closely related to the lifetime of clothing, and also plays a very important role in the health and well-being of humans. The care label inside the clothing indicates the recommended care operation, which usually contains multiple care symbols and multilingual textual descriptions. Repeated washing can lead to fading and deformation of labels. Care label recognition is a challenging task in the wild scene. In this paper, we propose a strong multi-modal multi-task baseline (abbreviated as MMFC), which combines image features and text features into a united framework. The Modality Mutual Transformation Module (MMTM) is employed to enhance the feature fusion. We refine the alignment of different modality features utilizing the methodology of contrastive learning and feature mapping. The lack of care label datasets has limited the development of intelligent care. Therefore, we introduce a new high-quality large-scale dataset called FashionCare, which has 30,477 images, a total of 157,907 fashion care symbols, six major categories, 66 subcategories and textual description. To our knowledge, this is the first large-scale dataset of care label. Extensive experiments on FashionCare show the effectiveness of MMFC. In order to demonstrate the few-shot recognition performance of MMFC, we build a sub-dataset called FashionCare-LT by constructing the tail subcategories. Both quantitative and qualitative results show that MMFC possesses exceptional few-shot recognition capabilities. We believe that FashionCare can also be further explored to benefit more fashion related tasks, such as the care analysis of different materials and fashion types. We also hope that FashionCare can serve as a new benchmark for large-scale fine-grained multimodal learning, and contribute to the development of multimodal recognition, understanding and analysis.