Understanding Intermediate Layers Using Linear Classifier Probes

AI Safety Fundamentals: Alignment - Een podcast door BlueDot Impact

Abstract:Neural network models have a reputation for being black boxes. We propose to monitor the features at every layer of a model and measure how suitable they are for classification. We use linear classifiers, which we refer to as "probes", trained entirely independently of the model itself. This helps us better understand the roles and dynamics of the intermediate layers. We demonstrate how this can be used to develop a better intuition about models and to diagnose potential problem...

Visit the podcast's native language site