Deep Learning-based RGB Image Modelling for Multi-component Pasture Biomass Estimation in Precision Grazing

V
Vijay H. Kalmani1,*
P
Prakash Ramesh Gadekar2
A
Amol C. Adamuthe3
1Department of Computer Science and Engineering, Kasegaon Education Society’s Rajarambapu Institute of Technology, Shivaji University, Sakharale-415 414, Maharashtra India.
2Marathwada Mitra Mandal’s College of Engineering, Pune-411 052, Maharashtra, India.
3Department of Information Technology, Kasegaon Education Society’s Rajarambapu Institute of Technology, Shivaji University, Sakharale-415 414, Maharashtra India.

Background: Accurate estimation of pasture biomass is significant for managing grazing, feed distribution and overall livestock productivity sustainably. The classical methods of cutting and weighing are the most trustworthy worst-case scenario but they are also foul slow, demanding much labour and are not useful for very large areas.

Methods: This research intended to construct a minimalist deep-learning framework that can accurately estimate various pasture biomass components from inexpensive RGB images and corresponding field data. The material used for the experiment included 1,785 tree-top images taken from Australian pastures, all linked to ground measurements of biomass fractions, vegetation indices and plant-height data. The proposed architecture brings together depth wise-separable convolutions, channel-attention mechanisms and Fourier feature projections to sink the spatial detail and capture the vegetation patterns. A stratified train–validation split and controlled augmentations were used to avoid data leakage and to boost the model’s robustness.

Result: The validation obtained very consistent accuracy across the five fractions of biomass with a weighted determination coefficient of 0.81. The developed framework is a sustainable, non-invasive method for estimating pasture biomass and thus it gives huge advantages in precision grazing and sustainable livestock management.

The global adoption of data-driven farming solutions has increased demand for improved biomass estimations. Traditional ground methods remain inefficient due to resource requirements. Deep learning, computer vision and RS provide scalable crop analysis capabilities (Akiva et al., 2022; Barriguinha et al., 2022; Maji et al., 2022). Deep learning models like CNN, CNN-GRU effectively model spectral and environmental factors (Wang et al., 2025; Lu et al., 2024). Analysis of multispectral images enables biomass estimation and yield prediction (Menon et al., 2025; Akcapınar and Apaydin, 2025). Remote Sensing provides non-destructive data, while vegetation indices integrated with deep learning deliver accurate results (Tian et al., 2020; Menon et al., 2025). AI-vision models enable precise biomass estimation (Padhiary et al., 2024). This integration is revolutionizing the domain by improving precision (Kamangir et al., 2024; Wang et al., 2025).
       
Still, the issue of biomass forecasting endures among different plants and climate types because of differences in model generalization. Accordingly, the author scrutinizes: (1) What are the ways to achieve the maximum performance from the lightest DL architectures for biomass prediction? (2) Do the frequency-enhanced mechanisms play a role in improving the stability of the models? (3) What are ways in which DL-RS systems can contribute to sustainable crop management? AgroVisionNet is a combination of depthwise-separable residual blocks and Fourier features for the purpose of dealing with spectral-spatial dependencies. The model has shown better performance by giving lower R2 error as compared to ResNet and CNN-GRU models when using CSIRO data. The study presents a fast biomass estimation model that is also effective in estimating crop biomass.
       
According to recent research, deep learning has proven to be very effective in agricultural imaging. VGG19 models have been used for the detection of plant diseases (Kashyap and Kashyap, 2025), while a combination of ResNet CNN and SVM has improved classification (Vidhya et al., 2025). The use of attention-enhanced architectures helps in the detection of diseases (Ashwini and Uma, 2025) and, at the same time, CNNs are able to detect wilting in soybeans (Na and Na, 2024). All these developments support the use of small models in agriculture. With the help of AI and ML, there has been a significant increase in the production of crops and agricultural GHG emissions have been reduced by 20% (Naeem et al., 2025; Avasthi et al., 2025). The obstacles encountered are mainly high expenses, the security of data and unavailability of properly trained staff. Among others, ML algorithms have greatly increased the accuracy of biomass estimation and predictive through multispectral data analysis (Menon et al., 2025; Akcapınar and Apaydin, 2025). The use of AI in aquaponic systems has been suggested for nutrient regulation purposes (Chandramenon et al., 2024). By using ML and DL models, forecasting of crop yields has become far more accurate than with traditional methods, while regression still applies for small-scale scenarios. The use of more sophisticated algorithms allows the prediction of the yields of wheat, maize and soybean by depending on the data from distant sensors and weather information (Johansen et al., 2020; Jabed and Masrah 2024; Saleem et al., 2023). The combination of AI and IoT sensors for resource management in agriculture leads to sustainable practices (Pawde and Dave 2025; Avasthi et al., 2025). The use of predictive analytics contributes to the improvement of the supply chain in terms of efficiency (Pallathadka et al., 2023). Key challenges include: the need for expensive, high-quality and accurately labeled pictures to train Deep Learning systems (Johansen et al., 2020; Abudu et al., 2025), limited model generalization over different environments (Zheng et al., 2025; Thamoonlest et al., 2025) and the opacity of black box models (Cheng et al., 2025; Wang and Yao, 2023). The lack of sufficient computing power also affects the use of Deep Learning in under-developed areas (Sangjan et al., 2023). Data access, model interpretability and infrastructure costs are the three major areas that future research should focus on to ensure that the implementation is sustainable (Zheng et al., 2025; Abudu et al., 2025; Moussaid et al., 2025).
       
The focus of this research work is on making the approach accurate, efficient and reproducible. The approach relies on RGB images and metadata. This approach doesn’t need expensive sensors. The approach relies on minimal processing with a lightweight architecture.
Dataset description
 
The experimental implementation and analysis were carried out at the Rajarambapu Institute of Technology, Department of Computer Science and Engineering, Rajaramnagar, Sangli, Maharashtra, during the 2025-2026 academic session. The dataset used in the present study was sourced from the CSIRO Biomass Prediction Challenge hosted on Kaggle (Liao et al., 2025), consisting of 1,785 field samples, where every sample was connected to an RGB image and the ground-measured biomass weights for the five components: Dry_Green_g, Dry_Dead_g, Dry_Clover_g, GDM_g and Dry_Total_g. Along with each sample, metadata are provided, which include image file path, NDVI values, average plant height, date of sampling and location information. Each .jpg file contains an image, which is assigned a common prefix identifier to connect it to a set of structured CSV files containing biomass data. An initial analysis of the data is conducted to examine data completeness, identify missing values, explore data for inconsistencies in values, as well as investigate data distribution for biomass values to ensure data readiness for regression modelling.
 
Data pre-processing
 
The entire pre-processing process aimed at obtaining clean input data and at eliminating the information sharing issue between the training and the validation samples. To make regression prediction less volatile, first, the extreme biomass records that went beyond the 99th percentile (111.76 g) were discovered and subsequently eliminated since these outliers greatly affected the updates of the gradients and thus the convergence of the model was skewed. The metadata table and the RGB image directory were then joined based on the common prefix that was extracted from each sample_id, thus generating a data frame containing image paths, biomass values, NDVI measurements and plant-height data. This integration ensured ensured that each training instance had both the visual and the tabular features needed for multimodal learning.
       
To deal with group leakage-where several records from the same physical plot could be included in both the training and the validation sets-the dataset was split with a prefix-based stratified split, making sure that all the biomass parts of one sample prefix were allotted only to either the training or validation subset. The tabular features NDVI and height were then normalised using min-max scaling according to Eq. (1) to preserve the comparable numerical ranges:

 
All normalization parameters related to NDVI and plant height were calculated only on the training subset and afterward, the same parameters were applied to the validation data without any changes, thus preventing the possibility of unintentional information leakage.
       
In the image processing pipeline, all the RGB pictures were downsized to (400 × 200) pixels, converted to PyTorch tensors and then normalised for each channel with ImageNet statistics as per Eq. (2):


Through this process, compatibility with the convolutional backbone was guaranteed and stable training was made possible. In addition to this, the training set underwent a mild augmentation that included random horizontal flipping and photometric jitter, which served to introduce the variability in illumination and canopy orientation in a controlled manner thus overfitting was diminished. The PyTorch torchvision.transforms library was used to carry out all the pre-processing steps and the eventual processed dataset was wrapped inside a custom Biomass Dataset loader to facilitate efficient training in batches.
 
Data augmentation
 
In order to make the model more robust against the natural variability of the pasture images, the controlled augmentation of data was applied to the training set. The photometric transformations were the means to eliminate the differences in light, shadow patterns and sensor exposures. The random brightness and contrast adjustments were applied by changing each pixel intensity (I) according to the Eq. (3):
 
 
Where,
(α) and (β)= The contrast and brightness factors respectively and they are randomly sampled within their predefined ranges.
       
By this operation, the model is exposed to various lighting conditions but the basic scene structure is still retained. Besides that, horizontal flipping was used as a form of geometric augmentation to mimic camera angle and pasture layout changes, thus lowering the model’s awareness of spatial bias. All the augmentations were done randomly during training using torchvision.transforms. The validation images stayed the same so that the evaluation was always the same.
 
Proposed model architecture
 
The prediction framework that was created in this research is determined by a tailored convolutional neural network (CNN) which is aimed at drawing out spatial, textural and structural clues from pasture canopy images while still being computationally efficient. The core of the system is based on a modified ResNet-style architecture where the regular convolutional layers are substituted with depthwise-separable convolutions, thus allowing the factorization of the filtering into space and channel-wise portions. This process not only decreases the number of learnable parameters but also increases the capacity of computation without losing any of the features being expressed, thus making the architecture fit for processing large-scale field images.
       
In order to improve the learning of channel-level representations, the use of Squeeze-and-Excitation (SE) blocks was applied after certain convolutional stages. These blocks adjust the features of the model by taking into account the relation of different channels, which in turn boosts the network’s ability to detect biologically significant signals like tonal gradients, leaf density and irregularities in pasture surface structure.
       
Besides the traditional convolutional features, the model also uses the Fourier feature projection, where the input image is changed through low-frequency sinusoidal encodings that are done before entering the regression head. This technique enables the network to sense the periodic and quasi-periodic vegetation patterns like tuft spacing, fine-scale grass texture and canopy oscillations, which would be tricky to show using spatial convolutions alone. The combination of the features obtained from the CNN and the Fourier-encoded representations result in a more powerful latent embedding for biomass estimation.
       
The final stage of this architecture is modeled by a multi-output regression head, which is denoted by fully connected layers, that predicts all five biomass constituents simultaneously. The transformation from the latent (X) to predicted values is expressed as in Eq. (4):
 
 
Where,
DG), (ŷDD), (ŷDC), (ŷGDM), (ŷDT)= The forecast values for Dry_Green_g, Dry_Dead_g, Dry_Clover_g, GDM_g and Dry_Total_g, in that order, respectively.
       
The schematic diagram of the proposed CNN-SE–Fourier model forward-propagation sequence is summarized in Algorithm 1. Every architectural element plays a role in the performance of the system and the roles are complementary. The depthwise-separable convolution lowers the computational requirement and at the same time maintains the spatial detail which is an advantage when deploying the model in limited-resource environments. The Squeeze-and-Excitation blocks improve the ability to differentiate the features among the channels which in turn increases the model’s sensitivity to the cues such as vegetation density and tonal variation that are of biological interest. Fourier feature projections allow capturing the functions in the frequency domain which aid in the capture of fine-scale texture and quasi-periodic canopy patterns that are hard to model with mere spatial convolutions due to their nature of being difficult to detect in spatially convoluted areas.

Algorithm 1: Forward Propagation of the Proposed CNN-SE-Fourier Biomass Estimation Model.
Input: RGB image (I).
Output: Predicted biomass vector (Ŷ).
 
Input normalization
 
(a) Resize the input image (I) to (400 × 200) pixels.
(b) Apply channel-wise normalization using Eq. (2):

 
Backbone feature extraction
 
(a) Pass (Inorm) through successive convolutional blocks.
(b) In each block, apply depthwise–separable convolution, batch normalization and ReLU activation.
(c) Apply Squeeze-and-Excitation (SE) channel recalibration in selected blocks.
(d) Denote the extracted CNN feature map as (Fcnn).
Fourier feature encoding
(a) Apply a low-frequency sinusoidal projection to the input image:
 
Ffourier = γ(I)
 
Where,
(γ (⋅) )= denotes the Fourier encoding function.
 
Latent feature fusion
 
(a) Concatenate the CNN and Fourier features to obtain the fused latent representation:
 
X = [;Ffourier].].
 
Multi-output regression head
 
(a) Pass the fused representation (X) through fully connected layers.
(b) Generate simultaneous predictions for all biomass components using Eq. (4):

 
Return
 
Return the predicted biomass vector (Ŷ).
 
Training configuration and optimization
 
Model optimization was done by applying the Smooth L1 loss function, also known as the Huber loss, which gives advantages of robustness against noisy biomass measurements and at the same time combines the two forms of penalties by using the quadratic and linear penalty regions. For the different biomass components, the loss was calculated based on equation (5):
 
 
When necessary, a weighted loss formulation was applied so as to give more importance to the agronomic importance of the individual biomass components, according to equation (6):
 

The optimization was done with the help of the AdamW optimizer which separates weight decay from gradient updates to enhance stability. The early learning rate and weight decay coefficient were set by trial and error and a ReduceLROnPlateau scheduler was used to lower the learning rate automatically when the validation performance stopped improving. Training was done in mini-batches and gradient clipping was active to avoid instability that might be caused by large gradient updates in the deeper layers. Also, dropout was used in the fully connected layers to fight overfitting.
               
All pre-processing, training and evaluation procedures were applied consistently across experiments to ensure methodological transparency and reproducibility.
The findings presented in this part are all drawn from the validation set of 20% that was held out, which was prepared by a prefix-based stratified split to prevent any data leakage. The analysis comprises the trends of training and validation losses, overall regression metrics, performance by component and predicted-versus-actual relationships, along with qualitative examples. This provides a comprehensive and neutral assessment of the model’s capability to measure pasture biomass under various field conditions.
 
Training and validation curves
 
Fig 1(a) shows stages of the training of the proposed model. The Smooth L1 loss decreases steadily over the epochs for training as well as validation data. This indicates that the model is training really well and is also able to optimize really well. The two lines also tend to be equal; they do not deviate much from each other, which indicates that the methods of preventing overfitting, including data augmentation, dropout, gradient clipping and using AdamW optimization, have really helped in preventing overfitting in this model because of the biomasses present in the data, as data in real-world images keeps changing, which leads to noise in ground truth values.

Fig 1: Training loss, predicted vs. actual biomass and residual error distribution for the proposed model.


       
For further evaluation of predictive performance, Fig 1(b) also shows a scatter plot of predicted versus measured biomass for all plant components together. The data points lie very close to the line of equality, which indicates that the model is well calibrated for both low and high biomass conditions. Deviations for a few high biomass samples are biologically reasonable as they account for natural variations in dense pasture vegetation.
       
Error distribution in Fig 1(c) is symmetrical, having an approximately central value of zero, which indicates that it does not introduce bias in either overestimation or underestimation. The data distribution is compact, having only a few data points as possible outliers, which further confirms validity in the results produced by it. From all the above analyses, it is clear that a stable model is developed for multi-output biomass estimation in pasture environments.
 
Quantitative performance evaluation
 
For assessing how well the CNN-SE-Fourier model predicted, regression analysis metrics were employed. The Mean Squared Error (MSE), Root Mean Squared Error (RMSE), Mean Absolute Error (MAE) and the weighted coefficient of determination  metrics, in this case, were employed for regression analysis. The calculations of the metrics for regression analysis were conducted using the validation split, which is set aside to provide an accurate indication of generalization for all five biomass fractions. The results of this analysis are presented in Table 1.

Table 1: Overall regression performance on the validation set.


       
The value of validation loss for the model came out to be 6.3722, whereas the total MSE was 124.4405, causing RMSE to be 11.159. This, in turn, shows that there was a strong predictive agreement with the biomass observations from the ground truth. The weighted  value of 0.8133 is an indication that the model explains more than 81% of the variance in biomass measurements which, in fact, is a proof of its capacity in capturing structural and spectral cues present in pasture images. The global MAE is different across components, due to different biomass scales, but the overall error behavior remains the same as that of field-level variability which is typically observed in pasture systems.
       
Introducing Smooth L1 (Huber) loss (Eq. 5) was a pivotal factor in the model’s characterization of being able to resist high-variance samples and measurement noise. It is the combination of the quadratic and linear error regions that makes the optimization process easier and prevents large errors from taking over the training signal, which is very crucial for the real-world biomass datasets that are diverse and noisy.
       
The consistency of results across biomass components and validation metrics indicates stable model behaviour rather than reliance on isolated favourable cases.
       
To evaluate the robustness of the proposed method thoroughly, we assessed its performance in comparison with standard baseline models, particularly with a mean predictor, linear regression and random forest regressor, all using the same validation split. Table 2 shows that the CNN-SE-Fourier model greatly exceeds all the baseline methods getting R² of 0.813 while random forest and linear regression receive 0.304 and 0.145 respectively. The huge reduction in RMSE and MAE implies that the proposed architecture is capable of capturing the spatial, textural and structural information found in pasture images which are not sufficiently represented by the traditional regression-based methods.

Table 2: Comparison of the proposed CNN-SE-fourier model with standard baseline regression models on the validation set.


 
Component-wise prediction
 
In addition to the overall performance reported in Section 3.2, a component-wise evaluation was performed to evaluate the model’s behavior for the five biomass outputs individually. The RMSE and MAE values obtained from the validation set for each biomass component are summarised in Table 3. These metrics are a clear indication of how the prediction accuracy varies among different components that are affected by dynamic range, texture complexity and visual separability.

Table 3: Component-wise RMSE and MAE for biomass prediction.


       
Due in part to its high variability, Dry_Green_g is considered a good proxy for pasture biomass. Therefore, only this component is subjected to detailed visual evaluation. Fig 2 shows the true values compared with predicted values for Dry_Green_g and it is clear that there is a strong alignment along the 1:1 line even though the green biomass in pasture systems is naturally wide spread. In Fig 3, quite a few representative images are shown, which based on their appearance and structure, are further proof of model’s capability of detecting and depicting the green vegetation cover.

Fig 2: Predicted versus true dry green g values with the 1:1 reference line (RMSE = 34.62 g).



Fig 3: Example dry green g image patches with corresponding true and predicted biomass values.


       
However, for the residues which comprise Dry_Dead_g, Dry_Clover_g, GDM_g and Dry_Total_g, the final numbers are still available in Table 2. The stability of the predictive behavior for these components is furthermore indicated by the consistent error magnitudes, which are in line with both the domain expectations and the inherent variation in ground-truth measurements.
 
Limitations
 
Even though the proposed model is generally strong, it still has certain limitations.
(i) The model’s ability to extract more than one characteristic from vegetation at a time, such as chlorophyll and moisture will be restricted by the sole use of RGB images. This can lead to a case of underestimation in high-biomass conditions.
(ii) The dataset is composed of a small number of extreme biomass samples, thereby restricting the overall applicability of the model apart from even the areas where Smooth L1 loss offers robustness.
(iii) Data from one geographic region might result in a model that is not able to adapt well to other areas or seasons with different pasture types and environmental conditions affecting the latter’s transferability.
(iv) RGB images of overlapping and dense canopies may be hard to visualize and thus, precise estimation can be sometimes difficult.
(v) The model yields outputs that are deterministic and there is no uncertainty quantification, hence, in operational settings, decision-making confidence may get affected.
 
Discussion and implications for precision agriculture
 
The study shows that robust estimation of pasture biomass is feasible using low-cost sensing systems and miniaturized deep learning systems even in practical field settings. The outcomes indicate that the available CNN–SE–Fourier model is the one that best predicts the biomass of different pasture condition and its reliability is for sure. The strong weighted  score (0.8133) and consistent component-wise performance suggest that the model does a very good job in spotting the visual cues of live and dead biomass, even in the case of the natural grasslands’ heterogeneity. These findings demonstrate robust performance within the evaluated pasture conditions and dataset. The higher variability of the predictions in the case of Dry_Green_g and Dry_Total_g indicate that they are the more complex and dynamic components of the whole process, while Dry_Clover_g and Dry Dead_g are the least variable and more stable ones with tighter error bounds.
       
The operational aspect of the model’s power to predict a number of biomass fractions from simple RGB imagery simply points to its great importance in the field of precision agriculture. Such a system can assist in the management of grazing by allowing quick and non-destructive assessments of feed availability, which in turn cuts down the need for labor-intensive field sampling. The lightweight architecture and stable learning behavior also render the approach suitable for low-cost devices, which can later be incorporated into mobile or UAV-based workflows. While the underlying design is not specific to a single environment, extending deployment across different regions, seasons, or pasture types will require additional validation.
       
In conclusion, this study highlights the potential of compact deep-learning models, which are augmented with attention as well as frequency features, to produce pasture insights that are ready for action. The incorporation of multi-spectral inputs or temporal data can take such systems to the level of being scalable. Such extensions are therefore identified as promising directions for future research rather than conclusions drawn from the present study.
A CNN-SE-Fourier architecture for multi-output pasture biomass prediction with RGB and metadata was presented as a very light-weight model in this study. Predicted performance of the model was really good with the weighted  of 0.8133 plus stable accuracy for the five biomass categories. The combination of depthwise separable convolutions, Squeeze-and-Excitation channel reweighting and Fourier feature injection played a major role in the strong feature extraction under different field conditions.
       
The findings assure that it is possible to get accurate estimates with very inexpensive image data, thus providing a good alternative to the traditional sampling method that is so labor-intensive. This means that the very crops that the farmers cannot allow go to waste because of their timely projected grazing can now be used and the pasture yield, moreover the non-destructive, scalable monitoring of vegetation dynamics, will not lose its value. Even though in the case of the extreme biomass values, slight underestimations happened, the overall performance indicates a strong potential for operational deployment.
       
Future enhancements could vary from the use of multispectral or temporal data, to exploring transformer-based encoders, to even creating uncertainty-aware models for prediction. These could help in sustainable pasture management by AI tools, not only in the short term but also in the longer term, by improving generalization capabilities from one region to another as well as from different seasons to different seasons.
 
Disclaimers
 
The views and conclusions expressed in this article are solely those of the authors and do not necessarily represent the views of their affiliated institutions. The authors are responsible for the accuracy and completeness of the information provided, but do not accept any liability for any direct or indirect losses resulting from the use of this content.
 
Informed consent
 
No animal or human subjects were involved in this study and therefore ethical approval and informed consent were not required.
The authors declare that there are no conflicts of interest regarding the publication of this article. No funding or sponsorship influenced the design of the study, data collection, analysis, decision to publish, or preparation of the manuscript.

  1. Abudu, D., Katie, C., Lucy, B. and Mirjam, R. (2025). A Meta-analysis of predictive accuracies and errors of biomass estimation models in Sub-Saharan Africa. Science of The Total Environment. 1001. doi: 10.1016/j.scitotenv.2025.180455.

  2. Akcapınar, M.C. and Halit, A. (2025). Early yield estimation in wheat using open access global datasets and artificial intelligence. Computers and Electronics in Agriculture. 237. doi: 10.1016/j.compag.2025.110486.

  3. Akiva, P., Benjamin, P., Aditi, R., Peter, O. and Kristin, D. (2022). Vision on the bog: Cranberry crop risk evaluation with deep learning. Computers and Electronics in Agriculture203. doi: 10.1016/j.compag.2022.107444. 

  4. Avasthi, S., Chauhan, R. and Suman, L.T. (2025). Artificial intelligence- powered agriculture and sustainable practices in developing countries. Hyperautomation in Precision Agriculture. pp 49-62.

  5. Ashwini, S.P. and Uma, S. (2025). A hybrid approach using ResNet50 and efficientNetB0 for attention-enhanced deep learning for early detection of apple plant diseases: A review. Agricultural Science Digest. 1-11. doi: 10.18805/ag.D-6375.

  6. Barriguinha, A., Jardim, B., de Castro, M.N. and Artur, G. (2022). Using NDVI, climate data and machine learning to estimate yield in the douro wine region. International Journal of Applied Earth Observation and Geoinformation. 114. doi: 10.1016/j.jag.2022.103069.

  7. Chandramenon, P., Aggoun, A. and Tchuenbou-Magaia, F. (2024). Smart approaches to aquaponics 4.0 with focus on water quality-Comprehensive review. Computers and Electronics in Agriculture. 225. doi: 10.1016/j.compag.2024. 109256.

  8. Cheng, F., Qiu, B., Yang, P., Wu, W., Yu, Q., Qian, J., Wu, B. and Chen, J. (2025). Crop sample prediction and early mapping based on historical data: Exploration of an explainable FKAN framework. Computers and Electronics in Agriculture. 237. doi: 10.1016/j.compag.2025.110689.

  9. Jabed, M.A. and Masrah, A.A.M. (2024). Crop yield prediction in agriculture: A comprehensive review of machine learning and deep learning approaches, with insights for future research and sustainability. Heliyon. 10. doi: 10.1016/ j.heliyon.2024.e40836.

  10. Johansen, K., Morton, M.J.L., Malbeteau, Y., Aragon, B., Al-Mashharawi, S. and Ziliani, M.G. et al. (2020). Predicting biomass and yield in a tomato phenotyping experiment using UAV imagery and random forest. Frontiers in Artificial Intelligence. 3. doi: 10.3389/frai.2020.00028.

  11. Kamangir, H., Sams, B.S., Dokoozlian, N., Sanchez, L. and Earles, J.M.  (2024). Large-scale spatio-temporal yield estimation via deep learning using satellite and management data fusion in vineyards. Computers and Electronics in Agriculture. 216. doi: 10.1016/j.compag.2023.108439.

  12. Kashyap, N. and Kashyap, A.K. (2025). Deep learning VGG19 model for precise plant disease detection. Agricultural Science Digest. 1-8. doi: 10.18805/ag.D-6220.

  13. Liao, Q., Wang, D., Pirie, R., Whelan, J., Haling, R., Liu, J., Khokher, R., Li, X., Plomecka, M. and Howard, A. (2025). CSIRO- Image 2 Biomass Prediction. https://kaggle.com/competitions/ csiro-biomass.

  14. Lu, J., Fu, H., Tang, X., Liu, Z., Huang, J., Zou, W., Chen, H., Sun, Y., Ning, X. and Jian, L. (2024). GOA-optimized deep learning for soybean yield estimation using multi-source remote sensing data. Scientific Reports. 14. doi: 10.1038/ s41598-024-57278-6.

  15. Maji, A.K., Marwaha, S., Kumar, S., Arora, A., Chinnusamy, V. and Shahnawazul, I. (2022). SlypNet: Spikelet-based yield prediction of wheat using advanced plant phenotyping and computer vision techniques. Frontiers in Plant Science. 13. doi: 10.3389/fpls.2022.889853.

  16. Menon, S., Aravinth J.A., Sankaran, R. and Kiran, P. (2025). Deep learning based farm-level crop yield prediction using multi-temporal satellite data for complex engineering application. Smart Agricultural Technology12. doi: 10.1016/j.atech.2025.101562.

  17. Moussaid, A., Gamoussi, Y. and Hamza, B. (2025). Hybrid CNN- LSTM model for predicting nitrogen, phosphorus and potassium (NPK) fertilization requirements: Integrating satellite spectral indices with field microclimate data. Internet of Things. 34. doi: 10.1016/j.iot.2025.101746.

  18. Naeem, M.Z., Jia, J., Wang, S., Poudel, S., Manjankattil, Y., Bailey, A.M. and Bourassa, D. (2025). Advancements in machine learning applications in poultry farming: A literature review. Journal of Applied Poultry Research. 34. doi: 10.1016/ j.japr.2025.100602.

  19. Na, M.H. and Na, I.S. (2024). Detection and classification of wilting in soybean crop using cutting-edge deep learning techniques.  Legume Research. 47(10): 1723-1729. doi: 10.18805/LRF-797.

  20. Padhiary, M., Saha, D., Kumar, R., Sethi, L.N. and Kumar, A.  (2024). Enhancing precision agriculture: A comprehensive review of machine learning and ai vision applications in all-terrain vehicle for farm automation. Smart Agricultural Technology. 8. doi: 10.1016/j.atech.2024.100483.

  21. Pallathadka, H., Mustafa, M., Sanchez, D.T., Sajja, G.S., Gour, S. and Mohd, N. (2023). Impact of machine learning on management, healthcare and agriculture. Materials Today: Proceedings. 80. doi: 10.1016/j.matpr.2021.07.042.

  22. Pawde, S. and Jaydeep, D. (2025). Applications of soft matter physics in food science: From molecular interactions to macro-scale food structures. Sustainable Food Technology3. doi: 10.1039/d5fb00172b.

  23. Saleem, S.R., Zaman, Q.U., Schumann, A.W. and Syed, M.Z.A.N. (2023). Variable rate technologies: Development, adaptation and opportunities in agriculture. Precision Agriculture. pp 103-122.

  24. Sangjan, W., McGee, R.J. and Sindhuja, S. (2023). Evaluation of forage quality in a pea breeding program using a hyperspectral sensing system. Computers and Electronics in Agriculture. 212. doi: 10.1016/j.compag.2023.108052.

  25. Thamoonlest, W., Posom, J., Saikaew, K., Phuphaphud, A., Chinapas, A., Panduangnat, L. and Khwantri, S. (2025). Forecasting gaps in sugarcane fields containing weeds using low- resolution UAV imagery based on a machine-learning approach. Smart Agricultural Technology. 10. doi: 10.1016/ j.atech.2025.100780.

  26. Tian, H., Wang, T., Liu, Y., Qiao, X. and Yanzhou, L. (2020). Computer vision technology in agricultural automation-A review. Information Processing in Agriculture. 7. doi: 10.1016/ j.inpa.2019.09.006.

  27. Vidhya, S.N.S., Rajarajeswari, S. and Basavaraj, G.N. (2025). Transfer learning-based areca nut (Areca catechu) disease detection using CNN and SVM approaches with ResNet-50 for improved deep learning performance. Indian Journal of Agricultural Research. 59(9): 1385- 1394. doi: 10.18805/IJARe.A-6404.

  28. Wang, H.S.H. and Yuan, Y. (2023). Machine learning for sustainable development and applications of biomass and biomass- derived carbonaceous materials in water and agricultural systems: A review. Resources, Conservation and Recycling. 190. doi: 10.1016/j.resconrec.2022.106847.

  29. Wang, Q., Zhang, Y., Ma, K.and Chenshuo, M. (2025). Prediction of long-term trends in biomass energy development suitability and optimization of feedstock collection layout based on deep learning algorithms. Journal of Cleaner Production. 495. doi: 10.1016/j.jclepro.2025.145079.

  30. Zheng, M., Liu, L., Wu, J. and Shuyi, D. (2025). Spatiotemporal evolution of agricultural greenhouses based on remote sensing data: A case study of weinan City in China. Journal of Agriculture and Food Research. 24. doi: 10.1016/j.jafr.2025.102326.

Deep Learning-based RGB Image Modelling for Multi-component Pasture Biomass Estimation in Precision Grazing

V
Vijay H. Kalmani1,*
P
Prakash Ramesh Gadekar2
A
Amol C. Adamuthe3
1Department of Computer Science and Engineering, Kasegaon Education Society’s Rajarambapu Institute of Technology, Shivaji University, Sakharale-415 414, Maharashtra India.
2Marathwada Mitra Mandal’s College of Engineering, Pune-411 052, Maharashtra, India.
3Department of Information Technology, Kasegaon Education Society’s Rajarambapu Institute of Technology, Shivaji University, Sakharale-415 414, Maharashtra India.

Background: Accurate estimation of pasture biomass is significant for managing grazing, feed distribution and overall livestock productivity sustainably. The classical methods of cutting and weighing are the most trustworthy worst-case scenario but they are also foul slow, demanding much labour and are not useful for very large areas.

Methods: This research intended to construct a minimalist deep-learning framework that can accurately estimate various pasture biomass components from inexpensive RGB images and corresponding field data. The material used for the experiment included 1,785 tree-top images taken from Australian pastures, all linked to ground measurements of biomass fractions, vegetation indices and plant-height data. The proposed architecture brings together depth wise-separable convolutions, channel-attention mechanisms and Fourier feature projections to sink the spatial detail and capture the vegetation patterns. A stratified train–validation split and controlled augmentations were used to avoid data leakage and to boost the model’s robustness.

Result: The validation obtained very consistent accuracy across the five fractions of biomass with a weighted determination coefficient of 0.81. The developed framework is a sustainable, non-invasive method for estimating pasture biomass and thus it gives huge advantages in precision grazing and sustainable livestock management.

The global adoption of data-driven farming solutions has increased demand for improved biomass estimations. Traditional ground methods remain inefficient due to resource requirements. Deep learning, computer vision and RS provide scalable crop analysis capabilities (Akiva et al., 2022; Barriguinha et al., 2022; Maji et al., 2022). Deep learning models like CNN, CNN-GRU effectively model spectral and environmental factors (Wang et al., 2025; Lu et al., 2024). Analysis of multispectral images enables biomass estimation and yield prediction (Menon et al., 2025; Akcapınar and Apaydin, 2025). Remote Sensing provides non-destructive data, while vegetation indices integrated with deep learning deliver accurate results (Tian et al., 2020; Menon et al., 2025). AI-vision models enable precise biomass estimation (Padhiary et al., 2024). This integration is revolutionizing the domain by improving precision (Kamangir et al., 2024; Wang et al., 2025).
       
Still, the issue of biomass forecasting endures among different plants and climate types because of differences in model generalization. Accordingly, the author scrutinizes: (1) What are the ways to achieve the maximum performance from the lightest DL architectures for biomass prediction? (2) Do the frequency-enhanced mechanisms play a role in improving the stability of the models? (3) What are ways in which DL-RS systems can contribute to sustainable crop management? AgroVisionNet is a combination of depthwise-separable residual blocks and Fourier features for the purpose of dealing with spectral-spatial dependencies. The model has shown better performance by giving lower R2 error as compared to ResNet and CNN-GRU models when using CSIRO data. The study presents a fast biomass estimation model that is also effective in estimating crop biomass.
       
According to recent research, deep learning has proven to be very effective in agricultural imaging. VGG19 models have been used for the detection of plant diseases (Kashyap and Kashyap, 2025), while a combination of ResNet CNN and SVM has improved classification (Vidhya et al., 2025). The use of attention-enhanced architectures helps in the detection of diseases (Ashwini and Uma, 2025) and, at the same time, CNNs are able to detect wilting in soybeans (Na and Na, 2024). All these developments support the use of small models in agriculture. With the help of AI and ML, there has been a significant increase in the production of crops and agricultural GHG emissions have been reduced by 20% (Naeem et al., 2025; Avasthi et al., 2025). The obstacles encountered are mainly high expenses, the security of data and unavailability of properly trained staff. Among others, ML algorithms have greatly increased the accuracy of biomass estimation and predictive through multispectral data analysis (Menon et al., 2025; Akcapınar and Apaydin, 2025). The use of AI in aquaponic systems has been suggested for nutrient regulation purposes (Chandramenon et al., 2024). By using ML and DL models, forecasting of crop yields has become far more accurate than with traditional methods, while regression still applies for small-scale scenarios. The use of more sophisticated algorithms allows the prediction of the yields of wheat, maize and soybean by depending on the data from distant sensors and weather information (Johansen et al., 2020; Jabed and Masrah 2024; Saleem et al., 2023). The combination of AI and IoT sensors for resource management in agriculture leads to sustainable practices (Pawde and Dave 2025; Avasthi et al., 2025). The use of predictive analytics contributes to the improvement of the supply chain in terms of efficiency (Pallathadka et al., 2023). Key challenges include: the need for expensive, high-quality and accurately labeled pictures to train Deep Learning systems (Johansen et al., 2020; Abudu et al., 2025), limited model generalization over different environments (Zheng et al., 2025; Thamoonlest et al., 2025) and the opacity of black box models (Cheng et al., 2025; Wang and Yao, 2023). The lack of sufficient computing power also affects the use of Deep Learning in under-developed areas (Sangjan et al., 2023). Data access, model interpretability and infrastructure costs are the three major areas that future research should focus on to ensure that the implementation is sustainable (Zheng et al., 2025; Abudu et al., 2025; Moussaid et al., 2025).
       
The focus of this research work is on making the approach accurate, efficient and reproducible. The approach relies on RGB images and metadata. This approach doesn’t need expensive sensors. The approach relies on minimal processing with a lightweight architecture.
Dataset description
 
The experimental implementation and analysis were carried out at the Rajarambapu Institute of Technology, Department of Computer Science and Engineering, Rajaramnagar, Sangli, Maharashtra, during the 2025-2026 academic session. The dataset used in the present study was sourced from the CSIRO Biomass Prediction Challenge hosted on Kaggle (Liao et al., 2025), consisting of 1,785 field samples, where every sample was connected to an RGB image and the ground-measured biomass weights for the five components: Dry_Green_g, Dry_Dead_g, Dry_Clover_g, GDM_g and Dry_Total_g. Along with each sample, metadata are provided, which include image file path, NDVI values, average plant height, date of sampling and location information. Each .jpg file contains an image, which is assigned a common prefix identifier to connect it to a set of structured CSV files containing biomass data. An initial analysis of the data is conducted to examine data completeness, identify missing values, explore data for inconsistencies in values, as well as investigate data distribution for biomass values to ensure data readiness for regression modelling.
 
Data pre-processing
 
The entire pre-processing process aimed at obtaining clean input data and at eliminating the information sharing issue between the training and the validation samples. To make regression prediction less volatile, first, the extreme biomass records that went beyond the 99th percentile (111.76 g) were discovered and subsequently eliminated since these outliers greatly affected the updates of the gradients and thus the convergence of the model was skewed. The metadata table and the RGB image directory were then joined based on the common prefix that was extracted from each sample_id, thus generating a data frame containing image paths, biomass values, NDVI measurements and plant-height data. This integration ensured ensured that each training instance had both the visual and the tabular features needed for multimodal learning.
       
To deal with group leakage-where several records from the same physical plot could be included in both the training and the validation sets-the dataset was split with a prefix-based stratified split, making sure that all the biomass parts of one sample prefix were allotted only to either the training or validation subset. The tabular features NDVI and height were then normalised using min-max scaling according to Eq. (1) to preserve the comparable numerical ranges:

 
All normalization parameters related to NDVI and plant height were calculated only on the training subset and afterward, the same parameters were applied to the validation data without any changes, thus preventing the possibility of unintentional information leakage.
       
In the image processing pipeline, all the RGB pictures were downsized to (400 × 200) pixels, converted to PyTorch tensors and then normalised for each channel with ImageNet statistics as per Eq. (2):


Through this process, compatibility with the convolutional backbone was guaranteed and stable training was made possible. In addition to this, the training set underwent a mild augmentation that included random horizontal flipping and photometric jitter, which served to introduce the variability in illumination and canopy orientation in a controlled manner thus overfitting was diminished. The PyTorch torchvision.transforms library was used to carry out all the pre-processing steps and the eventual processed dataset was wrapped inside a custom Biomass Dataset loader to facilitate efficient training in batches.
 
Data augmentation
 
In order to make the model more robust against the natural variability of the pasture images, the controlled augmentation of data was applied to the training set. The photometric transformations were the means to eliminate the differences in light, shadow patterns and sensor exposures. The random brightness and contrast adjustments were applied by changing each pixel intensity (I) according to the Eq. (3):
 
 
Where,
(α) and (β)= The contrast and brightness factors respectively and they are randomly sampled within their predefined ranges.
       
By this operation, the model is exposed to various lighting conditions but the basic scene structure is still retained. Besides that, horizontal flipping was used as a form of geometric augmentation to mimic camera angle and pasture layout changes, thus lowering the model’s awareness of spatial bias. All the augmentations were done randomly during training using torchvision.transforms. The validation images stayed the same so that the evaluation was always the same.
 
Proposed model architecture
 
The prediction framework that was created in this research is determined by a tailored convolutional neural network (CNN) which is aimed at drawing out spatial, textural and structural clues from pasture canopy images while still being computationally efficient. The core of the system is based on a modified ResNet-style architecture where the regular convolutional layers are substituted with depthwise-separable convolutions, thus allowing the factorization of the filtering into space and channel-wise portions. This process not only decreases the number of learnable parameters but also increases the capacity of computation without losing any of the features being expressed, thus making the architecture fit for processing large-scale field images.
       
In order to improve the learning of channel-level representations, the use of Squeeze-and-Excitation (SE) blocks was applied after certain convolutional stages. These blocks adjust the features of the model by taking into account the relation of different channels, which in turn boosts the network’s ability to detect biologically significant signals like tonal gradients, leaf density and irregularities in pasture surface structure.
       
Besides the traditional convolutional features, the model also uses the Fourier feature projection, where the input image is changed through low-frequency sinusoidal encodings that are done before entering the regression head. This technique enables the network to sense the periodic and quasi-periodic vegetation patterns like tuft spacing, fine-scale grass texture and canopy oscillations, which would be tricky to show using spatial convolutions alone. The combination of the features obtained from the CNN and the Fourier-encoded representations result in a more powerful latent embedding for biomass estimation.
       
The final stage of this architecture is modeled by a multi-output regression head, which is denoted by fully connected layers, that predicts all five biomass constituents simultaneously. The transformation from the latent (X) to predicted values is expressed as in Eq. (4):
 
 
Where,
DG), (ŷDD), (ŷDC), (ŷGDM), (ŷDT)= The forecast values for Dry_Green_g, Dry_Dead_g, Dry_Clover_g, GDM_g and Dry_Total_g, in that order, respectively.
       
The schematic diagram of the proposed CNN-SE–Fourier model forward-propagation sequence is summarized in Algorithm 1. Every architectural element plays a role in the performance of the system and the roles are complementary. The depthwise-separable convolution lowers the computational requirement and at the same time maintains the spatial detail which is an advantage when deploying the model in limited-resource environments. The Squeeze-and-Excitation blocks improve the ability to differentiate the features among the channels which in turn increases the model’s sensitivity to the cues such as vegetation density and tonal variation that are of biological interest. Fourier feature projections allow capturing the functions in the frequency domain which aid in the capture of fine-scale texture and quasi-periodic canopy patterns that are hard to model with mere spatial convolutions due to their nature of being difficult to detect in spatially convoluted areas.

Algorithm 1: Forward Propagation of the Proposed CNN-SE-Fourier Biomass Estimation Model.
Input: RGB image (I).
Output: Predicted biomass vector (Ŷ).
 
Input normalization
 
(a) Resize the input image (I) to (400 × 200) pixels.
(b) Apply channel-wise normalization using Eq. (2):

 
Backbone feature extraction
 
(a) Pass (Inorm) through successive convolutional blocks.
(b) In each block, apply depthwise–separable convolution, batch normalization and ReLU activation.
(c) Apply Squeeze-and-Excitation (SE) channel recalibration in selected blocks.
(d) Denote the extracted CNN feature map as (Fcnn).
Fourier feature encoding
(a) Apply a low-frequency sinusoidal projection to the input image:
 
Ffourier = γ(I)
 
Where,
(γ (⋅) )= denotes the Fourier encoding function.
 
Latent feature fusion
 
(a) Concatenate the CNN and Fourier features to obtain the fused latent representation:
 
X = [;Ffourier].].
 
Multi-output regression head
 
(a) Pass the fused representation (X) through fully connected layers.
(b) Generate simultaneous predictions for all biomass components using Eq. (4):

 
Return
 
Return the predicted biomass vector (Ŷ).
 
Training configuration and optimization
 
Model optimization was done by applying the Smooth L1 loss function, also known as the Huber loss, which gives advantages of robustness against noisy biomass measurements and at the same time combines the two forms of penalties by using the quadratic and linear penalty regions. For the different biomass components, the loss was calculated based on equation (5):
 
 
When necessary, a weighted loss formulation was applied so as to give more importance to the agronomic importance of the individual biomass components, according to equation (6):
 

The optimization was done with the help of the AdamW optimizer which separates weight decay from gradient updates to enhance stability. The early learning rate and weight decay coefficient were set by trial and error and a ReduceLROnPlateau scheduler was used to lower the learning rate automatically when the validation performance stopped improving. Training was done in mini-batches and gradient clipping was active to avoid instability that might be caused by large gradient updates in the deeper layers. Also, dropout was used in the fully connected layers to fight overfitting.
               
All pre-processing, training and evaluation procedures were applied consistently across experiments to ensure methodological transparency and reproducibility.
The findings presented in this part are all drawn from the validation set of 20% that was held out, which was prepared by a prefix-based stratified split to prevent any data leakage. The analysis comprises the trends of training and validation losses, overall regression metrics, performance by component and predicted-versus-actual relationships, along with qualitative examples. This provides a comprehensive and neutral assessment of the model’s capability to measure pasture biomass under various field conditions.
 
Training and validation curves
 
Fig 1(a) shows stages of the training of the proposed model. The Smooth L1 loss decreases steadily over the epochs for training as well as validation data. This indicates that the model is training really well and is also able to optimize really well. The two lines also tend to be equal; they do not deviate much from each other, which indicates that the methods of preventing overfitting, including data augmentation, dropout, gradient clipping and using AdamW optimization, have really helped in preventing overfitting in this model because of the biomasses present in the data, as data in real-world images keeps changing, which leads to noise in ground truth values.

Fig 1: Training loss, predicted vs. actual biomass and residual error distribution for the proposed model.


       
For further evaluation of predictive performance, Fig 1(b) also shows a scatter plot of predicted versus measured biomass for all plant components together. The data points lie very close to the line of equality, which indicates that the model is well calibrated for both low and high biomass conditions. Deviations for a few high biomass samples are biologically reasonable as they account for natural variations in dense pasture vegetation.
       
Error distribution in Fig 1(c) is symmetrical, having an approximately central value of zero, which indicates that it does not introduce bias in either overestimation or underestimation. The data distribution is compact, having only a few data points as possible outliers, which further confirms validity in the results produced by it. From all the above analyses, it is clear that a stable model is developed for multi-output biomass estimation in pasture environments.
 
Quantitative performance evaluation
 
For assessing how well the CNN-SE-Fourier model predicted, regression analysis metrics were employed. The Mean Squared Error (MSE), Root Mean Squared Error (RMSE), Mean Absolute Error (MAE) and the weighted coefficient of determination  metrics, in this case, were employed for regression analysis. The calculations of the metrics for regression analysis were conducted using the validation split, which is set aside to provide an accurate indication of generalization for all five biomass fractions. The results of this analysis are presented in Table 1.

Table 1: Overall regression performance on the validation set.


       
The value of validation loss for the model came out to be 6.3722, whereas the total MSE was 124.4405, causing RMSE to be 11.159. This, in turn, shows that there was a strong predictive agreement with the biomass observations from the ground truth. The weighted  value of 0.8133 is an indication that the model explains more than 81% of the variance in biomass measurements which, in fact, is a proof of its capacity in capturing structural and spectral cues present in pasture images. The global MAE is different across components, due to different biomass scales, but the overall error behavior remains the same as that of field-level variability which is typically observed in pasture systems.
       
Introducing Smooth L1 (Huber) loss (Eq. 5) was a pivotal factor in the model’s characterization of being able to resist high-variance samples and measurement noise. It is the combination of the quadratic and linear error regions that makes the optimization process easier and prevents large errors from taking over the training signal, which is very crucial for the real-world biomass datasets that are diverse and noisy.
       
The consistency of results across biomass components and validation metrics indicates stable model behaviour rather than reliance on isolated favourable cases.
       
To evaluate the robustness of the proposed method thoroughly, we assessed its performance in comparison with standard baseline models, particularly with a mean predictor, linear regression and random forest regressor, all using the same validation split. Table 2 shows that the CNN-SE-Fourier model greatly exceeds all the baseline methods getting R² of 0.813 while random forest and linear regression receive 0.304 and 0.145 respectively. The huge reduction in RMSE and MAE implies that the proposed architecture is capable of capturing the spatial, textural and structural information found in pasture images which are not sufficiently represented by the traditional regression-based methods.

Table 2: Comparison of the proposed CNN-SE-fourier model with standard baseline regression models on the validation set.


 
Component-wise prediction
 
In addition to the overall performance reported in Section 3.2, a component-wise evaluation was performed to evaluate the model’s behavior for the five biomass outputs individually. The RMSE and MAE values obtained from the validation set for each biomass component are summarised in Table 3. These metrics are a clear indication of how the prediction accuracy varies among different components that are affected by dynamic range, texture complexity and visual separability.

Table 3: Component-wise RMSE and MAE for biomass prediction.


       
Due in part to its high variability, Dry_Green_g is considered a good proxy for pasture biomass. Therefore, only this component is subjected to detailed visual evaluation. Fig 2 shows the true values compared with predicted values for Dry_Green_g and it is clear that there is a strong alignment along the 1:1 line even though the green biomass in pasture systems is naturally wide spread. In Fig 3, quite a few representative images are shown, which based on their appearance and structure, are further proof of model’s capability of detecting and depicting the green vegetation cover.

Fig 2: Predicted versus true dry green g values with the 1:1 reference line (RMSE = 34.62 g).



Fig 3: Example dry green g image patches with corresponding true and predicted biomass values.


       
However, for the residues which comprise Dry_Dead_g, Dry_Clover_g, GDM_g and Dry_Total_g, the final numbers are still available in Table 2. The stability of the predictive behavior for these components is furthermore indicated by the consistent error magnitudes, which are in line with both the domain expectations and the inherent variation in ground-truth measurements.
 
Limitations
 
Even though the proposed model is generally strong, it still has certain limitations.
(i) The model’s ability to extract more than one characteristic from vegetation at a time, such as chlorophyll and moisture will be restricted by the sole use of RGB images. This can lead to a case of underestimation in high-biomass conditions.
(ii) The dataset is composed of a small number of extreme biomass samples, thereby restricting the overall applicability of the model apart from even the areas where Smooth L1 loss offers robustness.
(iii) Data from one geographic region might result in a model that is not able to adapt well to other areas or seasons with different pasture types and environmental conditions affecting the latter’s transferability.
(iv) RGB images of overlapping and dense canopies may be hard to visualize and thus, precise estimation can be sometimes difficult.
(v) The model yields outputs that are deterministic and there is no uncertainty quantification, hence, in operational settings, decision-making confidence may get affected.
 
Discussion and implications for precision agriculture
 
The study shows that robust estimation of pasture biomass is feasible using low-cost sensing systems and miniaturized deep learning systems even in practical field settings. The outcomes indicate that the available CNN–SE–Fourier model is the one that best predicts the biomass of different pasture condition and its reliability is for sure. The strong weighted  score (0.8133) and consistent component-wise performance suggest that the model does a very good job in spotting the visual cues of live and dead biomass, even in the case of the natural grasslands’ heterogeneity. These findings demonstrate robust performance within the evaluated pasture conditions and dataset. The higher variability of the predictions in the case of Dry_Green_g and Dry_Total_g indicate that they are the more complex and dynamic components of the whole process, while Dry_Clover_g and Dry Dead_g are the least variable and more stable ones with tighter error bounds.
       
The operational aspect of the model’s power to predict a number of biomass fractions from simple RGB imagery simply points to its great importance in the field of precision agriculture. Such a system can assist in the management of grazing by allowing quick and non-destructive assessments of feed availability, which in turn cuts down the need for labor-intensive field sampling. The lightweight architecture and stable learning behavior also render the approach suitable for low-cost devices, which can later be incorporated into mobile or UAV-based workflows. While the underlying design is not specific to a single environment, extending deployment across different regions, seasons, or pasture types will require additional validation.
       
In conclusion, this study highlights the potential of compact deep-learning models, which are augmented with attention as well as frequency features, to produce pasture insights that are ready for action. The incorporation of multi-spectral inputs or temporal data can take such systems to the level of being scalable. Such extensions are therefore identified as promising directions for future research rather than conclusions drawn from the present study.
A CNN-SE-Fourier architecture for multi-output pasture biomass prediction with RGB and metadata was presented as a very light-weight model in this study. Predicted performance of the model was really good with the weighted  of 0.8133 plus stable accuracy for the five biomass categories. The combination of depthwise separable convolutions, Squeeze-and-Excitation channel reweighting and Fourier feature injection played a major role in the strong feature extraction under different field conditions.
       
The findings assure that it is possible to get accurate estimates with very inexpensive image data, thus providing a good alternative to the traditional sampling method that is so labor-intensive. This means that the very crops that the farmers cannot allow go to waste because of their timely projected grazing can now be used and the pasture yield, moreover the non-destructive, scalable monitoring of vegetation dynamics, will not lose its value. Even though in the case of the extreme biomass values, slight underestimations happened, the overall performance indicates a strong potential for operational deployment.
       
Future enhancements could vary from the use of multispectral or temporal data, to exploring transformer-based encoders, to even creating uncertainty-aware models for prediction. These could help in sustainable pasture management by AI tools, not only in the short term but also in the longer term, by improving generalization capabilities from one region to another as well as from different seasons to different seasons.
 
Disclaimers
 
The views and conclusions expressed in this article are solely those of the authors and do not necessarily represent the views of their affiliated institutions. The authors are responsible for the accuracy and completeness of the information provided, but do not accept any liability for any direct or indirect losses resulting from the use of this content.
 
Informed consent
 
No animal or human subjects were involved in this study and therefore ethical approval and informed consent were not required.
The authors declare that there are no conflicts of interest regarding the publication of this article. No funding or sponsorship influenced the design of the study, data collection, analysis, decision to publish, or preparation of the manuscript.

  1. Abudu, D., Katie, C., Lucy, B. and Mirjam, R. (2025). A Meta-analysis of predictive accuracies and errors of biomass estimation models in Sub-Saharan Africa. Science of The Total Environment. 1001. doi: 10.1016/j.scitotenv.2025.180455.

  2. Akcapınar, M.C. and Halit, A. (2025). Early yield estimation in wheat using open access global datasets and artificial intelligence. Computers and Electronics in Agriculture. 237. doi: 10.1016/j.compag.2025.110486.

  3. Akiva, P., Benjamin, P., Aditi, R., Peter, O. and Kristin, D. (2022). Vision on the bog: Cranberry crop risk evaluation with deep learning. Computers and Electronics in Agriculture203. doi: 10.1016/j.compag.2022.107444. 

  4. Avasthi, S., Chauhan, R. and Suman, L.T. (2025). Artificial intelligence- powered agriculture and sustainable practices in developing countries. Hyperautomation in Precision Agriculture. pp 49-62.

  5. Ashwini, S.P. and Uma, S. (2025). A hybrid approach using ResNet50 and efficientNetB0 for attention-enhanced deep learning for early detection of apple plant diseases: A review. Agricultural Science Digest. 1-11. doi: 10.18805/ag.D-6375.

  6. Barriguinha, A., Jardim, B., de Castro, M.N. and Artur, G. (2022). Using NDVI, climate data and machine learning to estimate yield in the douro wine region. International Journal of Applied Earth Observation and Geoinformation. 114. doi: 10.1016/j.jag.2022.103069.

  7. Chandramenon, P., Aggoun, A. and Tchuenbou-Magaia, F. (2024). Smart approaches to aquaponics 4.0 with focus on water quality-Comprehensive review. Computers and Electronics in Agriculture. 225. doi: 10.1016/j.compag.2024. 109256.

  8. Cheng, F., Qiu, B., Yang, P., Wu, W., Yu, Q., Qian, J., Wu, B. and Chen, J. (2025). Crop sample prediction and early mapping based on historical data: Exploration of an explainable FKAN framework. Computers and Electronics in Agriculture. 237. doi: 10.1016/j.compag.2025.110689.

  9. Jabed, M.A. and Masrah, A.A.M. (2024). Crop yield prediction in agriculture: A comprehensive review of machine learning and deep learning approaches, with insights for future research and sustainability. Heliyon. 10. doi: 10.1016/ j.heliyon.2024.e40836.

  10. Johansen, K., Morton, M.J.L., Malbeteau, Y., Aragon, B., Al-Mashharawi, S. and Ziliani, M.G. et al. (2020). Predicting biomass and yield in a tomato phenotyping experiment using UAV imagery and random forest. Frontiers in Artificial Intelligence. 3. doi: 10.3389/frai.2020.00028.

  11. Kamangir, H., Sams, B.S., Dokoozlian, N., Sanchez, L. and Earles, J.M.  (2024). Large-scale spatio-temporal yield estimation via deep learning using satellite and management data fusion in vineyards. Computers and Electronics in Agriculture. 216. doi: 10.1016/j.compag.2023.108439.

  12. Kashyap, N. and Kashyap, A.K. (2025). Deep learning VGG19 model for precise plant disease detection. Agricultural Science Digest. 1-8. doi: 10.18805/ag.D-6220.

  13. Liao, Q., Wang, D., Pirie, R., Whelan, J., Haling, R., Liu, J., Khokher, R., Li, X., Plomecka, M. and Howard, A. (2025). CSIRO- Image 2 Biomass Prediction. https://kaggle.com/competitions/ csiro-biomass.

  14. Lu, J., Fu, H., Tang, X., Liu, Z., Huang, J., Zou, W., Chen, H., Sun, Y., Ning, X. and Jian, L. (2024). GOA-optimized deep learning for soybean yield estimation using multi-source remote sensing data. Scientific Reports. 14. doi: 10.1038/ s41598-024-57278-6.

  15. Maji, A.K., Marwaha, S., Kumar, S., Arora, A., Chinnusamy, V. and Shahnawazul, I. (2022). SlypNet: Spikelet-based yield prediction of wheat using advanced plant phenotyping and computer vision techniques. Frontiers in Plant Science. 13. doi: 10.3389/fpls.2022.889853.

  16. Menon, S., Aravinth J.A., Sankaran, R. and Kiran, P. (2025). Deep learning based farm-level crop yield prediction using multi-temporal satellite data for complex engineering application. Smart Agricultural Technology12. doi: 10.1016/j.atech.2025.101562.

  17. Moussaid, A., Gamoussi, Y. and Hamza, B. (2025). Hybrid CNN- LSTM model for predicting nitrogen, phosphorus and potassium (NPK) fertilization requirements: Integrating satellite spectral indices with field microclimate data. Internet of Things. 34. doi: 10.1016/j.iot.2025.101746.

  18. Naeem, M.Z., Jia, J., Wang, S., Poudel, S., Manjankattil, Y., Bailey, A.M. and Bourassa, D. (2025). Advancements in machine learning applications in poultry farming: A literature review. Journal of Applied Poultry Research. 34. doi: 10.1016/ j.japr.2025.100602.

  19. Na, M.H. and Na, I.S. (2024). Detection and classification of wilting in soybean crop using cutting-edge deep learning techniques.  Legume Research. 47(10): 1723-1729. doi: 10.18805/LRF-797.

  20. Padhiary, M., Saha, D., Kumar, R., Sethi, L.N. and Kumar, A.  (2024). Enhancing precision agriculture: A comprehensive review of machine learning and ai vision applications in all-terrain vehicle for farm automation. Smart Agricultural Technology. 8. doi: 10.1016/j.atech.2024.100483.

  21. Pallathadka, H., Mustafa, M., Sanchez, D.T., Sajja, G.S., Gour, S. and Mohd, N. (2023). Impact of machine learning on management, healthcare and agriculture. Materials Today: Proceedings. 80. doi: 10.1016/j.matpr.2021.07.042.

  22. Pawde, S. and Jaydeep, D. (2025). Applications of soft matter physics in food science: From molecular interactions to macro-scale food structures. Sustainable Food Technology3. doi: 10.1039/d5fb00172b.

  23. Saleem, S.R., Zaman, Q.U., Schumann, A.W. and Syed, M.Z.A.N. (2023). Variable rate technologies: Development, adaptation and opportunities in agriculture. Precision Agriculture. pp 103-122.

  24. Sangjan, W., McGee, R.J. and Sindhuja, S. (2023). Evaluation of forage quality in a pea breeding program using a hyperspectral sensing system. Computers and Electronics in Agriculture. 212. doi: 10.1016/j.compag.2023.108052.

  25. Thamoonlest, W., Posom, J., Saikaew, K., Phuphaphud, A., Chinapas, A., Panduangnat, L. and Khwantri, S. (2025). Forecasting gaps in sugarcane fields containing weeds using low- resolution UAV imagery based on a machine-learning approach. Smart Agricultural Technology. 10. doi: 10.1016/ j.atech.2025.100780.

  26. Tian, H., Wang, T., Liu, Y., Qiao, X. and Yanzhou, L. (2020). Computer vision technology in agricultural automation-A review. Information Processing in Agriculture. 7. doi: 10.1016/ j.inpa.2019.09.006.

  27. Vidhya, S.N.S., Rajarajeswari, S. and Basavaraj, G.N. (2025). Transfer learning-based areca nut (Areca catechu) disease detection using CNN and SVM approaches with ResNet-50 for improved deep learning performance. Indian Journal of Agricultural Research. 59(9): 1385- 1394. doi: 10.18805/IJARe.A-6404.

  28. Wang, H.S.H. and Yuan, Y. (2023). Machine learning for sustainable development and applications of biomass and biomass- derived carbonaceous materials in water and agricultural systems: A review. Resources, Conservation and Recycling. 190. doi: 10.1016/j.resconrec.2022.106847.

  29. Wang, Q., Zhang, Y., Ma, K.and Chenshuo, M. (2025). Prediction of long-term trends in biomass energy development suitability and optimization of feedstock collection layout based on deep learning algorithms. Journal of Cleaner Production. 495. doi: 10.1016/j.jclepro.2025.145079.

  30. Zheng, M., Liu, L., Wu, J. and Shuyi, D. (2025). Spatiotemporal evolution of agricultural greenhouses based on remote sensing data: A case study of weinan City in China. Journal of Agriculture and Food Research. 24. doi: 10.1016/j.jafr.2025.102326.
In this Article
Published In
Agricultural Science Digest

Editorial Board

View all (0)