Paper Reading AI Learner

Telling Human and Machine Handwriting Apart

2026-01-16 18:45:16
Luis A. Leiva, Moises Diaz, Nuwan T. Attygalle, Miguel A. Ferrer, Rejean Plamondon

Abstract

Handwriting movements can be leveraged as a unique form of behavioral biometrics, to verify whether a real user is operating a device or application. This task can be framed as a reverse Turing test in which a computer has to detect if an input instance has been generated by a human or artificially. To tackle this task, we study ten public datasets of handwritten symbols (isolated characters, digits, gestures, pointing traces, and signatures) that are artificially reproduced using seven different synthesizers, including, among others, the Kinematic Theory (Sigma h model), generative adversarial networks, Transformers, and Diffusion models. We train a shallow recurrent neural network that achieves excellent performance (98.3 percent Area Under the ROC Curve (AUC) score and 1.4 percent equal error rate on average across all synthesizers and datasets) using nonfeaturized trajectory data as input. In few-shot settings, we show that our classifier achieves such an excellent performance when trained on just 10 percent of the data, as evaluated on the remaining 90% of the data as a test set. We further challenge our classifier in out-of-domain settings, and observe very competitive results as well. Our work has implications for computerized systems that need to verify human presence, and adds an additional layer of security to keep attackers at bay.

Abstract (translated)

手写动作可以被用作一种独特的行为生物识别方式,以验证实际用户是否在操作设备或应用程序。这一任务可以被视为一个反向图灵测试,在这种测试中,计算机需要检测输入实例是由真人生成的还是人工合成的。为了解决这个问题,我们研究了十个公开的手写符号数据集(包括孤立字符、数字、手势、指点轨迹和签名),这些数据集使用七种不同的生成器进行了人工再现,其中包括动学理论(Sigma h 模型)、生成对抗网络、Transformer 和扩散模型等。 我们训练了一个浅层循环神经网络,在不进行特征化处理的情况下以轨迹数据为输入,该网络在所有合成器和数据集中实现了卓越的性能(98.3% 的接收者操作特性曲线(ROC)面积得分和1.4% 的平均错误接受率)。此外,我们在少量样本设置下展示了当训练集仅为总数据量的10%,而剩余90%的数据作为测试集时,我们的分类器也能实现如此卓越的表现。我们进一步在跨域场景中挑战了这一分类器,并观察到了非常具有竞争力的结果。 这项工作对需要验证人类存在的计算机系统有着重要的影响,并为防止攻击者侵入添加了一层额外的安全保障。

URL

https://arxiv.org/abs/2601.11700

PDF

https://arxiv.org/pdf/2601.11700.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model LLM Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Robot Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Time_Series Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot