Automated production of batched unclonable micro-patterns anti-counterfeiting labels with strong robustness and rapid recognition speed
Abstract
Anti-counterfeiting technologies are indeed crucial for information security and protecting product authenticity. Traditional anti-counterfeiting methods have their limitations due to their clonable nature. Exploring new technologies, particularly those based on pixel-level textures, is a promising avenue to address the clonable issue due to high encoding capacity. However, research in this field is still in its early stages. This work introduces a new fluorescent anti-counterfeiting label technology with four key characteristics: efficient laser etching, high-throughput fabrication and segmentation, robustness aided by data augmentation, and an exceptionally high recognition speed. Specifically, the etching achieves a speed of 1,200 labels/3s, the high-throughput procedures yield a rate of 2,400 labels/4 min, and the total count is about 5.2 × 104 labels. The number of labels is further augmented to about 5.2 × 106 by implementing arbitrary rotation and brightness variation to enhance the robustness in the recognition procedure. We divide these labels into 44 categories based on differences in patterns. Utilizing machine learning methods, we have achieved a total recognition (including extraction and search process) time per label averaging 421.96 ms without classification, and 40.13 ms with classification. Specifically, the search process with classification is nearly fiftieth times shorter than the non-classification method, reaching
Keywords
INTRODUCTION
Under the temptation and drive of economic interests, the forgery of various commodities is increasing daily, which not only seriously harms the legitimate rights and interests of businesses and consumers, but also threatens national security, human health, and other aspects[1-3]. New and useful anti-counterfeiting materials and technologies have become an urgent need. People have explored various anti-counterfeiting strategies, including inkjet printing[4-7], bar code[7-12], optical anti-counterfeiting[13-16], watermarks[11,12,17,18], laser holograms[18-23], etc. Luminescence anti-counterfeiting, as one of the optical anti-counterfeiting technologies, derives from the changeable luminescence behaviors of luminescence materials under the alteration of various external stimuli such as excitation light[24-28], chemical reagent[26,27,29,30], heat[26,27,31], mechanical force[28,32], luminescence lifetime[33,34], etc. The luminescence anti-counterfeiting has been widely used in many areas such as renminbi (RMB), food safety, drug packaging,
Taking fluorescent anti-counterfeiting as an example, the mentioned multilevel anti-counterfeiting primarily relies on patterns for conveying anti-counterfeiting information, which inherently carries a limited amount of information. Compared to this approach, texture-based anti-counterfeiting technology operates at the pixel level, as the information used for anti-counterfeiting is small and is achieved through the randomness and uncertainty in the label generation process. This inherent unpredictability grants it natural unclonability. Therefore, texture-based anti-counterfeiting currently holds certain research value in the academic community. On the other hand, texture anti-counterfeiting often requires integration with machine learning methods due to the large amount of information that needs to be processed. He et al. designed and manufactured a physically non cloning multimode structured color anti-counterfeiting label based on artificial intelligence (AI) decodable amorphous photonic structures (APS) by introducing an intermediate white layer with water-induced transparency between the patterned APS layer and the black background layer[39]. By randomly rotating each of the six initial data to generate 500 images, 3,000 images were obtained to establish the database for deep learning. Samples taken by customers, which have different exposure, brightness, contrast, or a combination of these factors, were fed into the trained AI for validation. The results showed an accuracy for most of the authentic tags of > 0.99 indicating the high precision of the AI decoding process. It takes seconds to finish the whole authentication process. Liu et al. developed an inkjet-printable, AI decodable, unclonable, fluorescence security label[40]. About 3,000 images generated by randomly shifting and rotating six original images were provided for AI learning and classification. The deep learning model completed the entire authentication process for tags with different sharpness, brightness, rotation, amplification and mixing of these parameters, and it only takes a few seconds or less, and the matching degree for clear tags reaches about 99%. Lin et al. developed anti-counterfeiting labels with low reagent cost, simple and fast authentication, huge coding ability and high modifiability[41]. The scale of the labels in this work is smaller than those in the previous context. In the anti-counterfeiting verification, the two-step method was adopted. The software classified the shape of the patterns of a total of 4,000 images, and then verified the texture. This method improved the efficiency of image recognition. Through the shape-matching technology, the overall processing time (smartphone reading + data synchronization + authentication) is 12.17 s, and the matching time is 2.69 s. A total of 47 authentic images and 27 fake images were tested for verification, achieving accuracy rates of 97.9% and 100%, respectively.
There are current research flaws in pattern and texture anti-counterfeiting. The number of labels is small because there is no research on high-throughput synthesis, which is far from practical application. However, if the number of labels is increased by a few orders of magnitude, it will also put great pressure on subsequent batch processing of labels, construction of databases and follow-up recognition applications. The recognition speed using the current literature method is slow, and even in databases with only thousands of labels, it also requires a second-level recognition time.
This work utilizes a laser etching machine, adopts high-throughput preparation, and borrows the latest face recognition technology for etching, batch preparation, and rapid recognition of anti-counterfeiting labels. The etching process can achieve up to 1,200 patterns within just 3 s, and our high-throughput automated shooting speed has reached 2,400 labels/batch in 4 min. After using image segmentation methods and applying data augmentation techniques such as rotating angles and adjusting brightness, the initial set of 51,966 labels has been expanded to 5,196,600. By utilizing feature vector extraction method, a database has been quickly formed. Meanwhile, for a given label to be examined, we can achieve a recognition speed at the level of milliseconds, based on the database with a volume of 4,155,600, which is much faster than the previous literature. This recognition speed provides a rich margin for the recognition of larger data volume databases in the future. Our work has solved the problems of high-throughput etching, batch label generation, and rapid recognition for the practical application of texture anti-counterfeiting labels, making it a significant step forward in industrial applications.
MATERIALS AND METHODS
Preparation of micro-patterns of perovskite quantum dot film
In our work, we take 0.1468 g of PbBr2 (0.4 moles) and 0.0852 g of CsBr (0.4 moles), along with 1 mL of oleic acid and 0.5 mL of oleylamine, and thoroughly mix them with 10 mL of dimethylformamide. We use ultrasound agitation until the mixture becomes clear, resulting in the precursor solution for perovskite formation. Next, we heat 10 mL of toluene on a heating plate at 35 degrees Celsius for 10-15 min. Afterward, we carefully add 1 mL of the precursor solution to the toluene. We then stir the mixture thoroughly using a magnetic stirrer under constant temperature for 5 min to ensure uniform mixing. After mixing the obtained crude solution with ethyl acetate 1:2, we centrifuged it at 9,000 rpm for 3 min, removed the precipitate, and dispersed it into 2 mL of n-hexane to obtain yellow-green perovskite quantum dot solution. The prepared perovskite quantum dot solution was spin-coated onto the hydrophilic substrate after PLASMA treatment, and self-assembly was induced by heating and evaporation. After a heat process of
Synthesis of perovskite quantum dot nanocrystal film
Supplementary Figure 1 shows the X-ray diffraction (XRD) characterization data of the resulting sample. The main diffraction peaks at 15.0° (100), 21.2° (110), 30.2° (200), and 43.4° (220) were all characteristic peaks of CsPbBr3 quantum dot, indicating that it is a monoclinic phase. In addition, the fluorescence spectrum [Supplementary Figures 2 and 3] showed that CsPbBr3 quantum dot film had an absorption peak at 334 nm, a photoluminescence (PL) peak at 513 nm, and a stokes shift from 334 to 513 nm between emission and band-edge absorption. Furthermore, it can be seen from Supplementary Figure 4, that the short lifetime τ1 of the prepared CsPbBr3 quantum dot film was 10.0 ns, the long lifetime τ2 was 53.3 ns, the average lifetime was 44.9 ns, and the corresponding photoluminescence quantum yield (PLQY) was 89.4%. The high-resolution transmission electron microscopy image [Supplementary Figure 5] shows that the lattice fringes of perovskite quantum dots are 0.296 nm. The price of the precursor solution is relatively inexpensive, resulting in a lower cost, as shown in Supplementary Note 1.
The specific details of image segmentation methods
This method is divided into four steps in total. The first step is to convert the image into a grayscale image and use the Sobel operator to calculate the gradients in the x and y directions. Then, we subtract the gradients in the y direction from the x direction. Through this subtraction, we obtain an image region with high horizontal gradients and low vertical gradients. The second step is to remove noise from the image, which includes two parts. The first part is to use a low-pass filter to smooth the image (9 × 9 cores), which will help smooth out high-frequency noise in the image. The goal of a low-pass filter is to reduce the rate of change of the image. For example, replacing each pixel with the mean of the surrounding pixels can smooth and replace areas with significant intensity changes. The second part is binarization of blurred images. Any pixel in the gradient image that is not greater than 90 is set to 0 (black). Otherwise, pixels are set to 255 (white). The third step is to find all the contours of the image. Specifically, we use cv2.findContours() function to search for all contours, then calculate the area of all contours found, and finally select the top K contours based on the number of labels K in the image (K = 12). For contour detection model, we used the ResNet-18 network structure of convolution neural network. The fourth step is to calculate the minimum circumscribed matrix corresponding to the contour using the cv2.minAreaRect function, then crop the original image based on the coordinates provided by the circumscribed matrix, and finally store the cropped labels according to the contour or labels as needed.
The network architecture in the feather vector extraction model
The overall network architecture of ResNet-50, which is used to extract feature vectors from labels[42], is illustrated in Supplementary Figure 6, where the layers in each module add up to a total of 50 layers. Below the Bottleneck module in Supplementary Figure 6, the changes in the number of channels in the model are displayed. The specific structure of the Bottleneck is shown in Supplementary Figure 7.
RESULTS AND DISCUSSION
The whole process of anti-counterfeiting label production and recognition
The schematic diagram of the overall workflow of our work is shown in Figure 1. The entire workflow is segmented into three main stages: etching, label generation, and recognition, where label generation includes detailed steps of high-throughput photography, segmentation, and augmentation. Initially, the prepared thin films undergo etching with numerous pattern arrays. Subsequently, we used the homemade python program to control optical microscope for high-throughput and automatic characterization. After processing the captured photos using a segmentation model, we were able to quickly generate a large number of original labels. The 51,966 original labels were further augmented to 5,196,600 labels by adjusting various brightness and rotation angles. For the subsequent recognition process, we ultimately achieved a total time as fast as 40.1 ms and a minimum accuracy of 98.7%. The following are the details of each step.
Micro-patterns etching quantum dot film
First, we prepare perovskite quantum dot films and perform high-throughput etching to facilitate subsequent capturing and segmentation. The prepared solution is then spin-coated onto hydrophilic glass substrates treated with Plasma Cleaning Machine. The self-assembly is induced through heating and evaporation, and after constant heating at 75 degrees Celsius for 5 min on a hotplate, a perovskite quantum dot nanocrystal film is formed on the glass substrates. The formed crystal film has a unique micro-texture that exhibits physical unclonable function characteristics because of the randomness of the crystallization process[41]. The glass substrate with the film was then placed in the slot of the nanosecond laser engraving machine. After designing the desired pattern array using computer-aided design (CAD), the generated file was imported into the software controlling the nanosecond laser engraving machine to achieve high-throughput production. Through the software’s display module, we adjusted the focus and set alignment points, enabling the etching of micro-patterns and the borders, ensuring the alignment between the borders and the edges of the glass substrates. The etching process achieves a rate where nearly 1,200 labels can be manufactured in about 3 s. Our etching is for two purposes: (1) creating a bounding box for segmentation and (2) etching a pattern for subsequent image classification to accelerate recognition speed. It is notable that the output power of the laser is a key parameter to be adjusted. As shown in Figure 2A-D, it can be seen that the larger the etching power, the wider and darker the border. A distinct border is important for the image segmentation model in the later procedure. Therefore, the power of 128 mW [Figure 2D], which gives the most distinct border, was selected. If a higher laser power is employed, although it can make the boundaries clearer, thicker boundaries in the current label array will cause the edges of the labels to be connected with each other, making it impossible to separate. Therefore, 128 mW is the optimal laser power for this work.
Figure 2. (A-D) The fluorescence image of the generated micro-patterns film at 17, 38, 73, and 128 mW etching power (There is no obvious change in the etching area at 7 mW). The scale bar is 100 µm; (E) SEM images of a perovskite quantum dot film after etching micropatterns. The scale bar is 100 µm; (F) SEM images of etching lines of micropatterns. The scale bar is 20 µm; (G) SEM images of part of film. (G1) is the enlarged view of the non-etched area, (G2) is the enlarged view of the etched area. SEM: Scanning electron microscope.
Through CAD and nanosecond laser technology, it is possible to create extremely small and highly controllable labels. This is currently the most suitable technology for mass production in unclonable anti-counterfeiting labels due to the following reasons. Such a small size in this work is highly advantageous for generating larger data volumes in the later stages. The second reason is that a greater amount of valuable information is contained in the current labels. It is evident, as shown in Figure 2D, that the vast majority of the entire label consists of anti-counterfeiting texture information with an average content of about 75%, whereas previous work using brushing methods relied solely on the texture information applied through brushing[41]. The third reason is a significant improvement in the process of fabrication. The procedure in our work, including utilizing a spin coater for uniform coating and employing a laser engraving machine for automated etching, greatly reduces the need for manual labor, eliminating the requirement for a homemade coating device by hand[41]. The fourth reason is that the etching process achieves a fast rate where 1,200 labels can be manufactured in just 3 s. Furthermore, our method presents clear boundaries. The clarity of the boundaries obtained by laser engraving is significantly superior to those achieved by etching and subsequent brushing[41]. The scanning electron microscope (SEM) images are shown in Figure 2E-G. By comparing Figure 2G1 and G2, it can be seen that the grains in the etched area are more uniform and smaller than those in the non-etched area.
High-throughput photography and label segmentation methods
This is the crucial step in achieving batch experiments, and our automated photography method is as follows. As shown in Figure 1, during the process of high-throughput photography, in one batch, we have etched 200 images, each containing 12 labels. We imported the coordinates of the 200 images into the microscope’s automatic photography module, enabling the automated capturing. These coordinates were generated from the array arrangement designed during the CAD drawing, together with the physical location of the first image in the field of microscopy. A homemade Python program is utilized to generate the file containing all the coordinates simultaneously. Through the position information, the microscope automatically took 200 images; the number touches the uplimit of the instrument. As mentioned, each image is equipped with 12 labels, and this is attributed to two reasons: firstly, to maximize efficiency by fully utilizing the 1,920 × 1,440 resolution, and secondly, to prevent focusing issues. Through testing, having too many labels in one image can make focusing more challenging on the current microscopy, especially due to the unevenness of the samples. This method lays the foundation for automated production of anti-counterfeiting labels.
Due to the fact that each image (containing 12 labels) is less than 1 mm2, inevitably, the labels are not absolutely aligned within the image during the capture process, as shown in Figures 2 and 3; a proper label segmentation method is needed to fulfill the automated workflow. Automated image segmentation was unnecessary in previous work, due to a small number of labels[39-41]. However, it has become crucial in our high-throughput task. Additionally, given the relatively clear square contours, we can achieve the label segmentation effectively. This method processes images in four steps. First, the image is converted into a grayscale image. Second, the image is smoothed using a low-pass filter and then subjected to binarization to remove noise. Next, all contours are found, and the largest 12 contours are selected based on the number of labels in one image. Lastly, the minimum bounding rectangle for each contour is calculated, the original image is cropped accordingly, and the labeled images are stored as needed. We process all the images by the label segmentation method. Since there are 12 labels in each image and 200 images in each batch, about 2,400 labels were obtained after the label segmentation in about 40 s. Considering the 4 min of shooting, the overall high-throughput process of characterization takes about 5 min per batch, as shown in Figure 3. After repeating this process about 20 times, we finally obtained 51,966 original labels.
Data augmentation and classification
In order to boost the robustness of the machine learning model, we perform data augmentation. In practical applications, the image quality of anti-counterfeiting labels may be influenced by factors such as different lighting conditions and shooting angles. The significance of data augmentation lies in the substantial increase of the quantity of training data, and the robustness of the recognition process. By applying random rotations and brightness adjustments to a single label, we generate a total of 5,196,600 labels, i.e., 100 augmented ones per label. The rotation angle varies randomly from 0 to 360 degrees. The colors of the labels could be represented in the Hue, Saturation, Value (HSV) color space, where V (Value, ranging from 0 to 255) represents the brightness. The brightness adjustment in the data augmentation process involves adding the V component by an “x” value where “x” ∈ (-45, 90). The label examples after rotation angle and brightness adjustments (under a fluorescent microscope) are illustrated in Figure 4A. Figure 4B illustrates the brightness variations in the example shown in Figure 4A. It can be observed that the peak of the V-values for labels shifts towards higher values after brightening, while it shifts towards lower values after dimming the brightness.
Figure 4. (A) Examples after rotation angle and brightness adjustments (under a fluorescent microscope) and (B) histogram of corresponding brightness (V). The three lines represent the pixel distribution for each brightness value in the three images from (A), with each y-value corresponding to the number of pixels with a specific brightness value in the image.
The random and uncontrollable crystallization process will also cause the formed crystal film to have a unique texture[41]. In order to make the subsequent classification more convenient and rapid as the previous work[41], we etched several different types of patterns, including drawings, digits, letters, and Chinese characters, as shown in Supplementary Figure 8. There are about 44 groups of patterns in total. These patterns show the richness of etched patterns and a variety of patterns can be etched without being limited by whether the patterns are solid or not[41], and can classify 5,196,600 labels into 44 groups of texture database with different patterns. In the following recognition process, we selected 4,155,600 labels from the total 5,196,600 ones as the database (marked as real), with the remaining 1,041,000 labels used as fake ones (outside the database). We further explore the impact of the label classification on the recognition speed.
Overall recognition process
These are two steps in the recognition process: feature vector extraction and search in the database. For feature vector extraction, we utilized the ResNet-50 network structure of the convolutional neural network[42]. The loss function of this model is the Additive Angular Margin Loss (ArcFace), which significantly enhances discriminative power[43]. The method here is borrowed from a popular line of research in the field of face recognition, which involves adopting margins in the well-established softmax loss function to maximize class separability. The feature vectors of above-mentioned 4,155,600 labels have been extracted and stored into the database. For the search step of one particular label (inside or outside the database), we first augmented the model with a layer of multilayer perceptron, used for the classification of patterns if needed. By employing the Facebook AI Similarity Search (Faiss) retrieval framework, we were able to efficiently search the whole database to verify if the match was inside or not.
We selected 10,000 real labels from the database and fake labels not in the database to test the recognition speed and accuracy. The accuracy for real (fake) labels is determined by the percentage of real (fake) labels correctly identified as real (fake). The frequency distribution histogram is shown in Figure 5. We estimate based on the average value in the statistics. Figure 5A demonstrates that for real labels without pattern classification, the feature vector extraction time is approximately 30.70 ms per label in average, and the search time in the database is about 387.06 ms. Figure 5B shows that, without classification, the feature extraction time for fake labels is 30.93 ms, and the search time is 395.23 ms. When pattern classification is applied, as illustrated in Figure 5C and D, both real and fake labels show a slight increase in feature vector extraction time, i.e., 31.29 ms for real and 31.73 ms for fake. However, the search time significantly drops to 8.34 ms for real labels and 8.69 ms for fake labels, which is almost one-fiftieth of the search time for labels without classification. Therefore, for the unclassified method, the total times used for real and fake labels are respectively 417.76 and 426.16 ms. With classification, these times reduce to 39.83 ms for real and 40.42 ms for fake labels, indicating a significant efficiency gain. In terms of accuracy, a relatively high level was achieved. For the unclassified method, the accuracy in identifying real and fake labels was 99.77% and 98.7%, respectively. With the application of the classification method, these accuracies were 98.87% and 98.73%, respectively.
Figure 5. The frequency distribution histogram of (A) the feature vector extraction time and (B) search time for 10,000 real or fake labels with pattern classification and (C) the feature vector extraction time and (D) search time for 10,000 real or fake labels with pattern classification.
The demonstrated capability to process the recognition under a vast database (containing 4,155,600 labels) within a millisecond level shows exceptional efficiency. Such rapid processing is vital in practical scenarios where there is a need to swiftly handle a large number of labels. Furthermore, achieving at least a 98.7% accuracy rate in recognition is particularly impressive given the volume of the database. This high level of accuracy is crucial in anti-counterfeiting efforts, ensuring reliability and trust in the methods. Previous works[39-41] typically used data volumes of about 3,000 to 4,000, also achieving accuracy rates close to 98%, with processing times generally in seconds. Despite nearly similar accuracy rates, our work demonstrates a magnitude of improvement in both time and data volume compared to previous efforts. The implications of these findings are significant for future practical applications. They pave the way for the advanced recognition of unclonable anti-counterfeiting labels on a much larger scale, making the technology more applicable and relevant in real-world scenarios. This advancement not only enhances the capability of anti-counterfeiting technologies but also offers valuable insights and benchmarks for future developments in this field.
Since we are conducting pixel-level recognition and matching, reducing the etched area implies that the textured region will be larger. This increases the number of pixels we can identify, leading to improved accuracy. Meanwhile, the time used for etching could also be saved. In fact, we can even create labels without patterns to maximize the effective information content of each one, and still achieve an acceptable speed given the current volume of labels. The type of label without a pattern, as illustrated in Figure 6, offers the advantage of eliminating the need for CAD drawing, etching, and segmentation; it only needs cutting based on size as indicated by the dashed line in Figure 6. This maximizes the effective information. However, a drawback is that, due to the absence of classification, the identification process takes slightly longer as discussed above. In practical applications, cutting along virtual lines is required.
CONCLUSIONS
In summary, by optimizing the laser etching parameters, we obtain perovskite quantum dots nanocrystalline micro-patterns with micro size, high resolution, and different types, and the process achieves a rapid speed where 1,200 labels are produced in just 3 s. By designing a pattern array and inputting it into a python program to obtain files that can be read by the automatic photographing module of a microscope, we have achieved high-throughput photography, enabling 2,400 anti-counterfeiting labels to be shot within about 4 min. The image segmentation method divides these images into final anti-counterfeiting labels. And each label has a unique micro-texture. We generate a total of 5,196,600 labels by changing rotations and brightness randomly to a single label. By applying the ResNet-50 network for feature vector extraction and Faiss for search, the recognition process under millions of labels demonstrates strong advantages over literature in terms of efficiency (around 40 ms per label) and accuracy (≥ 98.7%), especially with classifications of patterns. For future data that may reach tens of millions or even billions in scale, our feature vector extraction time can still be maintained at around 30 ms. Additionally, pattern classification and/or a more efficient retrieval model are also helpful to improve the overall efficiency in the future.
DECLARATIONS
Acknowledgments
This work utilized the resources of Shanghai Engineering Research Center for Integrated Circuits and Advanced Display Materials, Shanghai Technical Service Center of Science and Engineering Computing, Shanghai University.
Authors’ contributions
Writing - review and editing, writing - original draft, visualization, validation, methodology, investigation, formal analysis, data curation, conceptualization: He Y
Writing - original draft, methodology, software, formal analysis, data curation, conceptualization: Zhang Z
Software, methodology: Xing Y
Methodology, investigation: Lang Z
Writing - review and editing, supervision, resources, project administration, methodology, investigation, funding acquisition, formal analysis, data curation, conceptualization: Wu J, Yang J
Availability of data and materials
Python source codes are available at https://github.com/7heyuzheng7/Pixel-level-Anti-counterfeiting-Labels-Recognition.
Financial support and sponsorship
This work was supported by the Natural Science Foundation of China (Grant Nos. 52172216 and 92163212) and the National Key Research and Development Program of China (No. 2021YFB3502200).
Conflicts of interest
All authors declared that there are no conflicts of interest.
Ethical approval and consent to participate
Not applicable.
Consent for publication
Not applicable.
Copyright
© The Author(s) 2024.
Supplementary Materials
REFERENCES
1. Zhang T, Wang L, Wang J, et al. Multimodal dynamic and unclonable anti-counterfeiting using robust diamond microparticles on heterogeneous substrate. Nat Commun 2023;14:2507.
2. Kumar V, Dottermusch S, Katumo N, Chauhan A, Richards BS, Howard IA. Unclonable anti-counterfeiting labels based on microlens arrays and luminescent microparticles. Adv Opt Mater 2022;10:2102402.
3. Cheng H, Wei X, Qiu H, Wang W, Su W, Zheng Y. Chemically stable fluorescent anti-counterfeiting labels achieved by UV-induced photolysis of nanocellulose. RSC Adv 2021;11:18381-6.
4. Ding L, Wang XD. Luminescent oxygen-sensitive ink to produce highly secured anticounterfeiting labels by inkjet printing. J Am Chem Soc 2020;142:13558-64.
5. You M, Zhong J, Hong Y, Duan Z, Lin M, Xu F. Inkjet printing of upconversion nanoparticles for anti-counterfeit applications. Nanoscale 2015;7:4423-31.
6. Zheng X, Zhu Y, Liu Y, et al. Inkjet-printed quantum dot fluorescent security labels with triple-level optical encryption. ACS Appl Mater Interfaces 2021;13:15701-8.
7. You M, Lin M, Wang S, et al. Three-dimensional quick response code based on inkjet printing of upconversion fluorescent nanoparticles for drug anti-counterfeiting. Nanoscale 2016;8:10096-104.
8. Wang T, Zheng H, You C, Ju J. A texture-hidden anti-counterfeiting QR code and authentication method. Sensors 2023;23:795.
9. Zhou Y, Zhao G, Bian J, et al. Multiplexed SERS barcodes for anti-counterfeiting. ACS Appl Mater Interfaces 2020;12:28532-8.
10. Chen R, Yu Y, Chen J, et al. Customized 2D barcode sensing for anti-counterfeiting application in smart IoT with fast encoding and information hiding. Sensors 2020;20:4926.
11. Xie R, Wu K, Xu G, Ouyang M. Research on anti-counterfeiting quick response 2D barcode techniques based on digital watermark. J Shanghai Jiaotong Univ 2013;18:443-7.
12. Xie R, Hong C, Zhu S, Tao D. Anti-counterfeiting digital watermarking algorithm for printed QR barcode. Neurocomputing 2015;167:625-35.
13. Hong W, Yuan Z, Chen X. Structural color materials for optical anticounterfeiting. Small 2020;16:e1907626.
14. Xie Y, Song Y, Sun G, Hu P, Bednarkiewicz A, Sun L. Lanthanide-doped heterostructured nanocomposites toward advanced optical anti-counterfeiting and information storage. Light Sci Appl 2022;11:150.
15. Gao D, Gao J, Gao F, et al. Quintuple-mode dynamic anti-counterfeiting using mult-mode persistent phosphors. J Mater Chem C 2021;9:16634-44.
16. Kuang Q, Hou X, Du C, Wang X, Gao D. Recent advances in the anti-counterfeiting applications of long persistent phosphors. Phys Chem Chem Phys 2023;25:17759-68.
17. Hu H, Zhong H, Chen C, Chen Q. Magnetically responsive photonic watermarks on banknotes. J Mater Chem C 2014;2:3695.
18. Wlodarczyk KL, Ardron M, Weston NJ, Hand DP. Holographic watermarks and steganographic markings for combating the counterfeiting practices of high-value metal products. J Mater Process Technol 2019;264:328-35.
19. Yetisen AK, Butt H, Mikulchyk T, et al. Color-selective 2.5D holograms on large-area flexible substrates for sensing and multilevel security. Adv Opt Mater 2016;4:1589-600.
20. Ruffato G, Rossi R, Massari M, Mafakheri E, Capaldo P, Romanato F. Design, fabrication and characterization of computer generated holograms for anti-counterfeiting applications using OAM beams as light decoders. Sci Rep 2017;7:18011.
21. Anastasiou A, Zacharaki EI, Tsakas A, Moustakas K, Alexandropoulos D. Laser fabrication and evaluation of holographic intrinsic physical unclonable functions. Sci Rep 2022;12:2891.
22. Tamulevičius T, Juodėnas M, Klinavičius T, et al. Dot-matrix hologram rendering algorithm and its validation through direct laser interference patterning. Sci Rep 2018;8:14245.
23. Miao J, Ding X, Zhou S, Gui C. Fabrication of dynamic holograms on polymer surface by direct laser writing for high-security anti-counterfeit applications. IEEE Access 2019;7:142926-33.
24. Sun T, Xu B, Chen B, et al. Anti-counterfeiting patterns encrypted with multi-mode luminescent nanotaggants. Nanoscale 2017;9:2701-5.
25. Li Z, Liu X, Wang G, et al. Photoresponsive supramolecular coordination polyelectrolyte as smart anticounterfeiting inks. Nat Commun 2021;12:1363.
26. de Azevedo LA, Gamonal A, Maier-queiroz R, et al. Multi-stimuli-responsive luminescent MCM48 hybrid for advanced anti-counterfeiting applications. J Mater Chem C 2021;9:9261-70.
27. Zhou H, Han J, Cuan J, Zhou Y. Responsive luminescent MOF materials for advanced anticounterfeiting. Chem Eng J 2022;431:134170.
28. He J, Yang Y, Li Y, et al. Multiple anti-counterfeiting guarantees from simple spiropyran derivatives with solid photochromism and mechanochromism. Cell Rep Phys Sci 2021;2:100643.
29. Yang H, Liu Y, Guo Z, et al. Hydrophobic carbon dots with blue dispersed emission and red aggregation-induced emission. Nat Commun 2019;10:1789.
30. Xia TF, Cao WQ, Cui YJ, Yang Y, Qian GD. Water-sensitive multicolor luminescence in lanthanide-organic framework for anti-counterfeiting. Opto Electron. Adv 2021;4:200063.
31. Qi Q, Li C, Liu X, et al. Solid-state photoinduced luminescence switch for advanced anticounterfeiting and super-resolution imaging applications. J Am Chem Soc 2017;139:16036-9.
32. Ma Z, Zhou J, Zhang J, et al. Mechanics-induced triple-mode anticounterfeiting and moving tactile sensing by simultaneously utilizing instantaneous and persistent mechanoluminescence. Mater Horiz 2019;6:2003-8.
33. Katumo N, Ruiz-preciado LA, Kumar V, Hernandez-sosa G, Richards BS, Howard IA. Anticounterfeiting labels with smartphone-readable dynamic luminescent patterns based on tailored persistent lifetimes in Gd2O2S:Eu3+/Ti4+. Adv Mater Technol 2021;6:2100047.
34. Huang Q, Mei X, Xie Z, et al. Photo-induced phosphorescence and mechanoluminescence switching in a simple purely organic molecule. J Mater Chem C 2019;7:2530-4.
35. Gao W, Ge W, Shi J, Tian Y, Zhu J, Li Y. Stretchable, flexible, and transparent SrAl2O4:Eu2+@TPU ultraviolet stimulated anti-counterfeiting film. Chem Eng J 2021;405:126949.
37. Yu X, Zhang H, Yu J. Luminescence anti-counterfeiting: from elementary to advanced. Aggregate 2021;2:20-34.
38. Bruno MDL, Lio GE, Ferraro A, et al. Flexible physical unclonable functions based on non-deterministically distributed dye-doped fibers and droplets. ACS Appl Mater Interfaces 2024;16:37063-72.
39. He X, Gu Y, Yu B, et al. Multi-mode structural-color anti-counterfeiting labels based on physically unclonable amorphous photonic structures with convenient artificial intelligence authentication. J Mater Chem C 2019;7:14069-74.
40. Liu Y, Han F, Li F, et al. Inkjet-printed unclonable quantum dot fluorescent anti-counterfeiting labels with artificial intelligence authentication. Nat Commun 2019;10:2409.
41. Lin Y, Zhang H, Feng J, et al. Unclonable micro-texture with clonable micro-shape towards rapid, convenient, and low-cost fluorescent anti-counterfeiting labels. Small 2021;17:e2100244.
42. He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR); 2016 Jun 27-30; Las Vegas, NV, USA. IEEE; 2016. p. 770-8.
Cite This Article
How to Cite
He, Y.; Zhang, Z.; Xing, Y.; Lang, Z.; Wu, J.; Yang, J. Automated production of batched unclonable micro-patterns anti-counterfeiting labels with strong robustness and rapid recognition speed. J. Mater. Inf. 2024, 4, 18. http://dx.doi.org/10.20517/jmi.2024.36
Download Citation
Export Citation File:
Type of Import
Tips on Downloading Citation
Citation Manager File Format
Type of Import
Direct Import: When the Direct Import option is selected (the default state), a dialogue box will give you the option to Save or Open the downloaded citation data. Choosing Open will either launch your citation manager or give you a choice of applications with which to use the metadata. The Save option saves the file locally for later use.
Indirect Import: When the Indirect Import option is selected, the metadata is displayed and may be copied and pasted as needed.
Comments
Comments must be written in English. Spam, offensive content, impersonation, and private information will not be permitted. If any comment is reported and identified as inappropriate content by OAE staff, the comment will be removed without notice. If you have any queries or need any help, please contact us at support@oaepublish.com.