Papers
Topics
Authors
Recent
Search
2000 character limit reached

ASCModule in ASCNet: Astronomical Image Classification

Updated 9 January 2026
  • ASCModule is a dedicated neural component that processes the luminance channel via parallel depthwise dilated convolutions with SE attention to capture subtle texture cues in astronomical images.
  • It integrates with a ResNet-34 RGB branch in ASCNet through a FusionBlock that aligns and concatenates features, enhancing overall classification performance.
  • Empirical studies reveal that ASCModule increases accuracy by around 15 percentage points over a ResNet-34 baseline, demonstrating its effectiveness in low-light, high-variability environments.

The ASCModule is a dedicated architectural component designed for fine-grained image classification, with specific application to astronomical all-sky camera (ASC) imagery. It forms a crucial part of ASCNet, a dual-branch neural network tailored for nighttime cloud type classification at astronomical observatories. The ASCModule processes the luminance (Y) channel of input images via parallel depthwise dilated convolutions with embedded lightweight Squeeze-and-Excitation (SE) attention. Its design enables capture of subtle luminance texture cues critical for classifying challenging phenomena in low-light and high-variability astronomical environments (Wang et al., 2 Jan 2026).

1. Architectural Placement and High-Level Workflow

ASCNet positions the ASCModule alongside a standard RGB processing branch, creating a complementary two-stream structure:

  • The RGB branch employs a ResNet-34 backbone for extracting broad semantic features across color channels.
  • The luminance branch processes the grayscale Y channel using the ASCModule, which is optimized for fine-scale cloud texture discrimination.

Given a batch of color images XRB×3×H×WX \in \mathbb{R}^{B \times 3 \times H \times W}, the network computes luminance as

Y=0.299XR+0.587XG+0.114XB,YRB×1×H×W.Y = 0.299 X_R + 0.587 X_G + 0.114 X_B,\qquad Y \in \mathbb{R}^{B \times 1 \times H \times W}.

The subsequent feature maps FrgbF_{rgb} (from ResNet-34) and FyF_{y} (from the ASCModule) are aligned in spatial resolution, concatenated to form FcatF_{cat}, and processed by a FusionBlock before passing to pooling and the classifier.

2. Internal Structure of ASCModule

The ASCModule consists of k=3k=3 parallel branches, each parameterized by a different dilation rate (d1=1d_1=1, d2=2d_2=2, d3=3d_3=3). Each branch executes the following sequence:

a) Depthwise Dilated Convolution Block

  • Input: U(0)=YU^{(0)} = Y, shape (B,1,H,W)(B,1,H,W)
  • Operation: Apply depthwise 3×\times3 convolution, dilation dd

Ub,c,i,j(1)=u=13v=13Kc,u,v(d)Ub,c,i+d(u2),j+d(v2)(0).U^{(1)}_{b,c,i,j} = \sum_{u=1}^3\sum_{v=1}^3 K^{(d)}_{c,u,v} \cdot U^{(0)}_{b,c, i + d(u-2),\, j + d(v-2)}.

  • Batch normalization: U(2)=BN(U(1))U^{(2)} = \mathrm{BN}(U^{(1)})
  • Pointwise (1×\times1) convolution: U(3)=Conv1×1(U(2))U^{(3)} = \mathrm{Conv}_{1 \times 1}(U^{(2)}), expands to (B,C,H,W)(B,C',H',W')

b) Squeeze-and-Excitation Attention Block

  • Squeeze: Global average pooling per channel,

zc=1HWi=1Hj=1WUc,i,j(3)z_c = \frac{1}{H'W'} \sum_{i=1}^{H'} \sum_{j=1}^{W'} U^{(3)}_{c,i,j}

  • Excitation: Two fully-connected layers,

s=σ ⁣(W2δ(W1z))s = \sigma\!\left(W_2\,\delta(W_1z)\right)

where W1RC/r×CW_1 \in \mathbb{R}^{C'/r \times C'}, W2RC×C/rW_2 \in \mathbb{R}^{C' \times C'/r}, rr is SE reduction ratio, σ\sigma is sigmoid, δ\delta is ReLU.

  • Channel-wise scaling: F=sU(3)F = s \odot U^{(3)}

Each branch outputs FiRB×C×H×WF_i \in \mathbb{R}^{B \times C' \times H' \times W'}; the final ASCModule output is

Fy=Concat(F1,F2,F3)    RB×(3C)×H×W.F_{y} = \mathrm{Concat}(F_1, F_2, F_3)\;\in\; \mathbb{R}^{B \times (3C') \times H' \times W'}.

3. Channel Dimensions and Feature Fusion

The design selects CC' such that the combined channel dimension from all luminance branches (Cy=3CC_{y} = 3C') approximates the channel count in the ResNet RGB branch (Crgb=512C_{rgb}=512), enabling direct concatenation. For C=170C'=170, Cy510C_{y} \approx 510.

Fusion is implemented as follows:

  1. Spatial alignment: ResNet feature map and FyF_{y} are reshaped to equal (H,W)(H', W').
  2. Channel concatenation: Fcat=Concat(Frgb,Fy)RB×(Crgb+Cy)×H×WF_{cat} = \mathrm{Concat}(F_{rgb}, F_{y}) \in \mathbb{R}^{B\times (C_{rgb}+C_y)\times H'\times W'}.
  3. FusionBlock: 1×\times1 channel reduction, depthwise separable convolution with residual connection, Efficient Channel Attention (ECA), and dropout for regularization.

4. Computational Complexity

Parameter and FLOP cost is minimal relative to standard multi-channel convolutions:

  • Per branch: $9$ weights (depthwise 3×\times3), CC' (pointwise), 2C2/r2C'^2/r (SE FC layers).
  • For C=170C'=170, r=16r=16: 9+170+2×1702/1637919+170+2\times 170^2/16 \approx 3791 weights per branch; total for three branches: 11\approx 11K.
  • FLOPs: Depthwise conv: 3×3×1×HW3\times 3\times 1 \times H'W'; pointwise: 1×C×HW1\times C'\times H'W' per branch.
  • A plausible implication is that the ASCModule can be integrated into deep CNNs with negligible overhead relative to standard residual blocks.

Relative to a standard ResNet-34 residual block with C=512C=512, which has 4.7\sim4.7M parameters, the ASCModule increases model capacity and representational richness at small cost.

5. Ablation Results and Performance Gains

Empirical studies show a substantial increase in classification metrics when the ASCModule is combined with the FusionBlock and ResNet-34 backbone. The following table summarizes the ablation study ((Wang et al., 2 Jan 2026), Table 4):

Scheme Accuracy (%) Precision (%) Recall (%) F1 Score (%)
ResNet34 (baseline) 77.82 50.76 55.53 51.89
+ FusionBlock 89.94 77.86 78.09 74.25
+ ASCModule (ASCNet) 92.66 83.26 84.25 83.67

This demonstrates a relative gain of approximately 15 percentage points in accuracy when employing both the ASCModule and FusionBlock above the ResNet-34-only baseline.

6. Implementation Details and Pseudocode

The primary logic of the ASCModule and its role in ASCNet is as follows:

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
def ASCModule(Y, C_prime=170, reduction=16):
    branch_outputs = []
    for d in [1,2,3]:
        U = depthwise_conv2d(Y, kernel_size=3, dilation=d, padding=d)  # (B,1,H',W')
        U = BatchNorm(U)
        U = conv2d(U, out_channels=C_prime, kernel_size=1)             # (B,C',H',W')
        z = global_avg_pool(U)                                         # (B,C')
        e = relu( dense(z, out_features=C_prime//reduction) )          # (B, C'/r)
        s = sigmoid( dense(e, out_features=C_prime) )                  # (B, C')
        s = s.view(B, C_prime, 1,1)
        Fi = s * U                                                     # (B,C',H',W')
        branch_outputs.append(Fi)
    Fy = concatenate(branch_outputs, dim=1)                            # (B,3*C',H',W')
    return Fy 

def ASCNet_forward(X):
    Y = 0.299*X[:,0] + 0.587*X[:,1] + 0.114*X[:,2]      # (B,H,W)
    Y = Y.unsqueeze(1)                                  # (B,1,H,W)
    Frgb = ResNet34_backbone(X)                         # (B,512,H',W')
    Fy   = ASCModule(Y, C_prime=170)                    # (B,510,H',W')
    Fcat = concatenate([Frgb, Fy], dim=1)               # (B,1022,H',W')
    Ffus = FusionBlock(Fcat)
    out = global_avg_pool(Ffus)
    out = Flatten(out)
    out = BatchNorm1d(out)
    logits = Linear(out, out_features=5)
    return softmax(logits)

7. Significance and Application Context

Integration of the ASCModule into ASCNet results in significantly improved performance on the classification of nighttime all-sky camera images, as measured on data from the Muztagh-ata site. The approach leverages the luminance channel to extract discriminative features not directly accessible in the RGB data pathway. Emphasis on fine-grained texture via multiscale dilation and SE recalibration yields high accuracy (92.66%), precision (83.26%), recall (84.25%), and F1 score (83.67%) as validated against human-labeled standards. The modularity and computational parsimony of the ASCModule suggests potential for broader application in other domains requiring luminance-focused, texture-sensitive analysis (Wang et al., 2 Jan 2026).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to ASCModule.