[3] Sengupta, S., et al. (2020). Background matting v2. CVPR .
Single GPU, version v1.2 of AIarty Matting, no trimap support (AIarty does not accept trimaps). Future work should test on video sequences and VR applications. References [1] Qin, X., et al. (2020). U²-Net: Going deeper with nested U-structure for salient object detection. Pattern Recognition , 106, 107404. aiarty matting
Author: [Your Name/Institution] Date: [Current Date] Abstract Image matting—the task of accurately extracting foreground elements with fine boundary details—remains a challenge for conventional computer vision methods, particularly for hair, fur, and translucent objects. This paper evaluates AIarty Matting , an AI-driven solution that leverages generative neural networks to produce alpha mattes. Using a dataset of 500 diverse images (portraits, e-commerce products, nature scenes), we compare AIarty Matting against three established methods: U²-Net, MODNet, and Adobe Photoshop’s “Select Subject” (AI-based). Metrics include SAD (Sum of Absolute Differences), MSE (Mean Squared Error), inference time per image, and user-rated boundary quality. Results indicate that AIarty Matting outperforms MODNet in fine detail retention (SAD improvement of 12.4%) but requires 1.8× higher inference latency. We conclude with recommendations for optimizing generative matting for real-time applications. [3] Sengupta, S
Table 2: Designers’ rating of edge realism and artifact absence. References [1] Qin, X
Image matting, generative AI, alpha matte, edge detection, AIarty 1. Introduction Image matting is essential for photo editing, film compositing, and augmented reality. Traditional methods (e.g., GrabCut, Closed-Form Matting) require user-supplied trimaps or scribbles. Recent deep learning approaches have enabled automatic matting, but they struggle with complex boundaries or low-contrast regions.
[4] AIarty Matting User Guide (v1.2). Hypothetical documentation, 2025.
