ASCModule in ASCNet: Astronomical Image Classification
- ASCModule is a dedicated neural component that processes the luminance channel via parallel depthwise dilated convolutions with SE attention to capture subtle texture cues in astronomical images.
- It integrates with a ResNet-34 RGB branch in ASCNet through a FusionBlock that aligns and concatenates features, enhancing overall classification performance.
- Empirical studies reveal that ASCModule increases accuracy by around 15 percentage points over a ResNet-34 baseline, demonstrating its effectiveness in low-light, high-variability environments.
The ASCModule is a dedicated architectural component designed for fine-grained image classification, with specific application to astronomical all-sky camera (ASC) imagery. It forms a crucial part of ASCNet, a dual-branch neural network tailored for nighttime cloud type classification at astronomical observatories. The ASCModule processes the luminance (Y) channel of input images via parallel depthwise dilated convolutions with embedded lightweight Squeeze-and-Excitation (SE) attention. Its design enables capture of subtle luminance texture cues critical for classifying challenging phenomena in low-light and high-variability astronomical environments (Wang et al., 2 Jan 2026).
1. Architectural Placement and High-Level Workflow
ASCNet positions the ASCModule alongside a standard RGB processing branch, creating a complementary two-stream structure:
- The RGB branch employs a ResNet-34 backbone for extracting broad semantic features across color channels.
- The luminance branch processes the grayscale Y channel using the ASCModule, which is optimized for fine-scale cloud texture discrimination.
Given a batch of color images , the network computes luminance as
The subsequent feature maps (from ResNet-34) and (from the ASCModule) are aligned in spatial resolution, concatenated to form , and processed by a FusionBlock before passing to pooling and the classifier.
2. Internal Structure of ASCModule
The ASCModule consists of parallel branches, each parameterized by a different dilation rate (, , ). Each branch executes the following sequence:
a) Depthwise Dilated Convolution Block
- Input: , shape
- Operation: Apply depthwise 33 convolution, dilation
- Batch normalization:
- Pointwise (11) convolution: , expands to
b) Squeeze-and-Excitation Attention Block
- Squeeze: Global average pooling per channel,
- Excitation: Two fully-connected layers,
where , , is SE reduction ratio, is sigmoid, is ReLU.
- Channel-wise scaling:
Each branch outputs ; the final ASCModule output is
3. Channel Dimensions and Feature Fusion
The design selects such that the combined channel dimension from all luminance branches () approximates the channel count in the ResNet RGB branch (), enabling direct concatenation. For , .
Fusion is implemented as follows:
- Spatial alignment: ResNet feature map and are reshaped to equal .
- Channel concatenation: .
- FusionBlock: 11 channel reduction, depthwise separable convolution with residual connection, Efficient Channel Attention (ECA), and dropout for regularization.
4. Computational Complexity
Parameter and FLOP cost is minimal relative to standard multi-channel convolutions:
- Per branch: $9$ weights (depthwise 33), (pointwise), (SE FC layers).
- For , : weights per branch; total for three branches: K.
- FLOPs: Depthwise conv: ; pointwise: per branch.
- A plausible implication is that the ASCModule can be integrated into deep CNNs with negligible overhead relative to standard residual blocks.
Relative to a standard ResNet-34 residual block with , which has M parameters, the ASCModule increases model capacity and representational richness at small cost.
5. Ablation Results and Performance Gains
Empirical studies show a substantial increase in classification metrics when the ASCModule is combined with the FusionBlock and ResNet-34 backbone. The following table summarizes the ablation study ((Wang et al., 2 Jan 2026), Table 4):
| Scheme | Accuracy (%) | Precision (%) | Recall (%) | F1 Score (%) |
|---|---|---|---|---|
| ResNet34 (baseline) | 77.82 | 50.76 | 55.53 | 51.89 |
| + FusionBlock | 89.94 | 77.86 | 78.09 | 74.25 |
| + ASCModule (ASCNet) | 92.66 | 83.26 | 84.25 | 83.67 |
This demonstrates a relative gain of approximately 15 percentage points in accuracy when employing both the ASCModule and FusionBlock above the ResNet-34-only baseline.
6. Implementation Details and Pseudocode
The primary logic of the ASCModule and its role in ASCNet is as follows:
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 |
def ASCModule(Y, C_prime=170, reduction=16): branch_outputs = [] for d in [1,2,3]: U = depthwise_conv2d(Y, kernel_size=3, dilation=d, padding=d) # (B,1,H',W') U = BatchNorm(U) U = conv2d(U, out_channels=C_prime, kernel_size=1) # (B,C',H',W') z = global_avg_pool(U) # (B,C') e = relu( dense(z, out_features=C_prime//reduction) ) # (B, C'/r) s = sigmoid( dense(e, out_features=C_prime) ) # (B, C') s = s.view(B, C_prime, 1,1) Fi = s * U # (B,C',H',W') branch_outputs.append(Fi) Fy = concatenate(branch_outputs, dim=1) # (B,3*C',H',W') return Fy def ASCNet_forward(X): Y = 0.299*X[:,0] + 0.587*X[:,1] + 0.114*X[:,2] # (B,H,W) Y = Y.unsqueeze(1) # (B,1,H,W) Frgb = ResNet34_backbone(X) # (B,512,H',W') Fy = ASCModule(Y, C_prime=170) # (B,510,H',W') Fcat = concatenate([Frgb, Fy], dim=1) # (B,1022,H',W') Ffus = FusionBlock(Fcat) out = global_avg_pool(Ffus) out = Flatten(out) out = BatchNorm1d(out) logits = Linear(out, out_features=5) return softmax(logits) |
7. Significance and Application Context
Integration of the ASCModule into ASCNet results in significantly improved performance on the classification of nighttime all-sky camera images, as measured on data from the Muztagh-ata site. The approach leverages the luminance channel to extract discriminative features not directly accessible in the RGB data pathway. Emphasis on fine-grained texture via multiscale dilation and SE recalibration yields high accuracy (92.66%), precision (83.26%), recall (84.25%), and F1 score (83.67%) as validated against human-labeled standards. The modularity and computational parsimony of the ASCModule suggests potential for broader application in other domains requiring luminance-focused, texture-sensitive analysis (Wang et al., 2 Jan 2026).