Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ml 85% Match Research Paper AI researchers focused on interpretability,Computer vision engineers,Machine learning practitioners,Data scientists 1 month ago

A Methodology for Transparent Logic-Based Classification Using a Multi-Task Convolutional Tsetlin Machine

computer-vision › diffusion-models
📄 Abstract

Abstract: The Tsetlin Machine (TM) is a novel machine learning paradigm that employs finite-state automata for learning and utilizes propositional logic to represent patterns. Due to its simplistic approach, TMs are inherently more interpretable than learning algorithms based on Neural Networks. The Convolutional TM has shown comparable performance on various datasets such as MNIST, K-MNIST, F-MNIST and CIFAR-2. In this paper, we explore the applicability of the TM architecture for large-scale multi-channel (RGB) image classification. We propose a methodology to generate both local interpretations and global class representations. The local interpretations can be used to explain the model predictions while the global class representations aggregate important patterns for each class. These interpretations summarize the knowledge captured by the convolutional clauses, which can be visualized as images. We evaluate our methods on MNIST and CelebA datasets, using models that achieve 98.5\% accuracy on MNIST and 86.56\% F1-score on CelebA (compared to 88.07\% for ResNet50) respectively. We show that the TM performs competitively to this deep learning model while maintaining its interpretability, even in large-scale complex training environments. This contributes to a better understanding of TM clauses and provides insights into how these models can be applied to more complex and diverse datasets.

Key Contributions

This paper proposes a methodology for transparent logic-based classification using a Multi-Task Convolutional Tsetlin Machine (CTM). It enables the generation of both local and global interpretations, visualizing the learned logic clauses as images, and achieves high accuracy on benchmark datasets.

Business Value

Offers highly interpretable AI models for image classification tasks, crucial in domains requiring transparency and accountability, such as medical diagnostics or autonomous systems.

View Code on GitHub