Simple Summary This study presents a novel cross-modal fusion method for accurate recognition of fish feeding intensity in complex underwater environments. Using acoustic and visual data from hydrophones and cameras, we develop a two-stage attention mechanism that adaptively combines complementary information from both modalities to overcome the limitations of single-modal approaches. The first stage enhances individual modal representations through cross-modal interactions, while the second stage dynamically adjusts fusion weights based on environmental conditions and modal reliability. Experimental results demonstrate that our method significantly outperforms existing single-modal and conventional fusion approaches, achieving superior accuracy in underwater scenarios. This work provides robust technical support for intelligent aquaculture monitoring and precision fish farming management.
ABSTRACT
PUBLICATION RECORD
- Publication year
2025
- Venue
Animals
- Publication date
2025-07-31
- Fields of study
Medicine, Computer Science, Environmental Science
- Identifiers
- External record
- Source metadata
Semantic Scholar, PubMed
CITATION MAP
EXTRACTION MAP
CLAIMS
- No claims are published for this paper.
CONCEPTS
- No concepts are published for this paper.
REFERENCES
Showing 1-37 of 37 references · Page 1 of 1
CITED BY
Showing 1-4 of 4 citing papers · Page 1 of 1