2.2. Image Preprocessing and Contour Extraction for Bowl-Type Tableware
Bowl contours are typically composed of straight segments and circular arcs. Because contour extraction and capacity estimation require fitting these geometric elements, it is essential to accurately obtain contour edge coordinates to ensure reliable dimensional and geometric characterization. In image processing, filtering is commonly used for denoising, feature enhancement, and contour smoothing; typical approaches include Gaussian, median, and mean filtering. Gaussian filtering performs a weighted average, effectively suppressing high-frequency random noise and smoothing fine details, but it may weaken edges[
9]. Median filtering is a nonlinear technique that replaces each pixel with the median value in its neighborhood; it is particularly effective against salt-and-pepper noise and better preserves edges[
10]. Mean filtering is computationally efficient and suitable for uniformly distributed noise, but it tends to blur edges and remove fine details[
11]. For the frontal-view and top-down bowl images, we applied all three filters and compared their ability to preserve contour details. The results indicate that median filtering achieves effective denoising while introducing less blur and fewer artifacts, thereby retaining bowl edge information more completely. Accordingly, we adopt median filtering in the proposed pipeline to provide clearer and more accurate inputs for subsequent contour segmentation, feature extraction, and capacity prediction.
Figure 3 shows bowl images processed using the three filtering methods.
After filtering the frontal-view and top-down images, we extract the inner and outer rim contours from the top-down view to support subsequent bowl wall-thickness prediction. To this end, we design a two-stage pipeline consisting of (i) image processing and (ii) contour extraction. To evaluate the effectiveness of different method combinations, we tested 40 combinations (each comprising one image-processing method and one contour-extraction method). By visually inspecting and comparing the resulting images, we selected the optimal scheme for subsequent bowl wall-thickness prediction.
In stage (i), we apply seven common image-processing methods to enhance bowl contrast and improve contour separability. (a): CLAHE performs contrast-limited adaptive histogram equalization in local regions, enhancing low-contrast details and emphasizing bowl edges and specular highlights[
12]. (b): Histogram Equalization adjusts the global intensity distribution to alleviate uneven background illumination[
13]. (c): Inversion reverses pixel intensities to increase the visual distinction between the bowl and the background when their contrast is weak[
14]. In morphological processing. (d): Top Hat subtracts the opened image from the original image to enhance bright details and suppress slowly varying background components, thereby highlighting contour structures[
15]. (e): Black Hat subtracts the original image from the closed image to enhance dark-region details, which is beneficial when the bowl’s inner surface appears darker[
16]. (f): Closing Operation applies dilation followed by erosion to fill small gaps and holes, improving contour continuity. (g): Opening Operation applies erosion followed by dilation to remove small artifacts and isolated noise, producing cleaner images for subsequent contour extraction[
17].
In stage (ii), we employ four edge-detection techniques for contour extraction. (a): High-Pass Filter preserves high-frequency components to accentuate intensity transitions, thereby sharpening the bowl-rim contours[
18]. (b): Canny applies a multi-step procedure—including denoising, gradient computation, non-maximum suppression, and hysteresis thresholding—to stably extract the primary outer contours and edges while suppressing noise[
19]. (c): The Laplace localizes edges using the second-order grayscale derivative; although it is sensitive to noise, it provides complementary capability for capturing inner-wall details[
20]. (d): Morphological Gradient extracts contours by computing the difference between dilation and erosion, which is well suited to bowls with regular structures and distinct edges, highlighting the outer boundary while reducing interference from internal texture[
21].
By combining the stage-one image processing methods with the stage-two contour extraction methods, we processed the top-down images of bowls and obtained image-processing results for 40 different method combinations. The experimental results show that, as illustrated in
Figure 4, the combination of CLAHE and Canny significantly enhances the difference between the inner and outer walls, outperforming the other methods as well as the original image. Under this setting, both high-pass filter and Canny can extract the inner- and outer-wall contours; however, high-pass filtering is prone to adhesion. In contrast, Canny produces fewer breakpoints and branches, yields smoother contours, and provides clearer separation, making the processed top-down bowl images more suitable for subsequent bowl thickness prediction.
2.4. Mask Segmentation and Volume Estimation of Bowls
In this subsection, we use the frontal-view images paired with the bowl top-down images as input and adopt Mask R-CNN as the base framework. Since each image contains only a single bowl, the primary goal is to obtain a high-quality mask with smooth boundaries and accurate details[
30]. However, the standard Mask R-CNN, which is designed for multi-object and multi-class detection, introduces redundant computation and may provide insufficient contour delineation for our task. To address this issue, we build an improved Mask R-CNN framework and compare three backbone networks. We further propose a geometry-aware feature pyramid structure, GFPN (Geometry-aware Feature Pyramid Network), tailored to this study, and investigate its effect on bowl mask segmentation. Finally, we select the model with the highest segmentation accuracy and extract keypoints from the predicted bowl mask to infer the rim diameter, base diameter, bowl height, and the sidewall contour corrected by the predicted wall thickness, which are then used to estimate bowl capacity.
2.4.1. Improved Mask R-CNN framework
In this subsection, we propose an improved Mask R-CNN baseline that is optimized for the characteristics of the mask segmentation task. Conventional Mask R-CNN is primarily designed for multi-object, multi-class detection and must jointly support classification and detection, which leads to substantial computational overhead on classification and insufficient delineation of fine bowl contours. To address these issues, we retain the two-stage architecture and the mask branch, while improving the feature extraction and pyramid fusion components. For region proposal generation, we continue to use the Region Proposal Network (RPN). However, considering the single-bowl image setting and the requirement for high geometric precision, we simplify the RPN structure and its output configuration[
31]. At each scale, we keep only a small set of scale and aspect-ratio combinations that better match bowl contours, thereby reducing the number of proposals and improving convergence efficiency. Standard Mask R-CNN produces hundreds of proposals and then filters them via NMS before further processing; in our task, this is redundant and may introduce uncertainty. Therefore, after NMS we retain only 50–100 high-confidence proposals and feed them into RoI Align and the segmentation head for subsequent processing[
32]. In the RoI head, we perform single-class classification without distinguishing specific bowl types. This avoids allocating additional fully connected layers and classification losses for fine-grained categorization. Meanwhile, we keep the original loss functions, which helps the feature representations in the mask and geometric branches remain more focused, thereby improving contour segmentation accuracy.
2.4.2. Three Different Backbone Networks
For the backbone, we adopt three architectures—Res Net-50, Res NeXt-50 (32×4d), and Res Net-C4—to maintain compatibility with the standard Mask R-CNN and to facilitate controlled comparisons. The reasons for selecting these three backbones are as follows:
Res Net-50 serves as the baseline for comparison. By introducing residual blocks and skip connections, it effectively alleviates the vanishing-gradient problem in deep networks[
33]
Res NeXt-50 (32×4d) can be regarded as augmenting the Res Net-50 bottleneck with 32 parallel 3×3 convolutional groups, each with 4 channels, implemented via grouped convolutions, thereby improving texture representation and contour modeling capability[
34]
Res Net-C4 retains only C1–C4 as the shared backbone and does not further down sample to C5, trading some high-level semantic information for higher spatial resolution and finer local structure, which benefits precise localization of geometric boundaries such as the bowl rim and base[
35]
The architectures of the three backbones are illustrated in the upper-left part of Figure 8. The three boxes correspond to the three backbone variants. Beneath C2, using C2 as an example, the figure further compares the convolutional operation in Res Net-50 with standard convolutions against that in Res NeXt-50 (32×4d) with grouped convolutions. By comparing these three backbones, we can analyze how backbone architecture affects mask quality and geometric fitting accuracy.
2.4.3. Geometry-aware Feature Pyramid Network
The Feature Pyramid Network (FPN) is a multi-scale architecture widely used in object detection. It enhances the recognition of targets at different scales by fusing feature maps from multiple hierarchical levels in a bottom-up pyramid manner. In our task, FPN plays the following roles[
35]:
Improved contour segmentation accuracy: Because bowl contours may exhibit subtle variations, FPN effectively combines fine-grained details from shallow layers with high-level semantic information from deeper layers, thereby improving the accuracy of bowl contour segmentation. In particular, for segmenting the bowl rim and base, FPN strengthens the model’s sensitivity to fine details, ensuring accurate extraction of circular contours
Reduced computational redundancy: In conventional Mask R-CNN, processing multi-scale features can introduce redundant computation. FPN mitigates this issue through efficient feature fusion, improving computational efficiency. This advantage is especially important for bowl segmentation tasks that require real-time performance or high-throughput processing
Multi-scale feature extraction: FPN extracts features at different hierarchical levels and fuses them, enabling the model to recognize targets at various scales within the image
To further emphasize bowl contour features, we propose a Geometry-aware Feature Pyramid Network (GFPN) that includes only P3–P5, built on the C3–C5 outputs of Res Net-50 and Res NeXt-50 (32×4d). This pyramid follows the standard top-down design with convolutions, up sampling, and convolutions. On this basis, we introduce a Geometry-aware Prior Modulation (GPM) module and the CBAM (Convolutional Block Attention Module) attention mechanism. With multi-scale feature enhancement, GFPN strengthens boundary responses and thereby improves the accuracy of bowl contour segmentation.
GPM enhances the model’s understanding of geometric structures by incorporating geometric prior knowledge. In mask segmentation, GPM modulates the feature maps according to the geometric shape information in the input image, thereby improving sensitivity to geometric boundaries such as bowl contours[
36]. CBAM strengthens feature representation through two components: channel attention and spatial attention. Channel attention models the importance of each channel and selectively amplifies informative channel features, whereas spatial attention emphasizes different spatial locations to increase the model’s responses to salient regions[
37]. GFPN is derived from the original FPN with targeted modifications. The overall architecture of GFPN is shown in
Figure 7, and the procedure is described as follows:
The frontal-view bowl image is first converted to grayscale, and the resulting grayscale image is denoted as
. The corresponding morphological gradient image is defined in Eq. (7):
In this formulation,
takes larger values near the bowl contour and is close to zero in background regions, where
and
denote the height and width of the original image. To inject scale-matched geometric priors into each FPN level,
is downsampled at multiple scales and encoded via convolution. Taking the third level with a stride of 8 as an example, average pooling is first applied to obtain a coarse-scale boundary map
, which is then encoded using a
convolution to produce the geometric prior feature
. The details are given in Eqs. (8) and (9):
In these equations,
denotes the coarse-scale boundary map at the third level,
represents the average pooling operation, and
and
.
In these equations, denotes the geometric prior feature corresponding to a stride of 8. is a function symbol representing the convolution operation, which transforms the input map into a new feature map . denotes a convolution.
Similarly, for the levels with strides of 16 and 32, the corresponding geometric prior features and can be obtained. Thus, for the three levels with strides of 8, 16, and 32, we derive the geometric prior features , , and . These features are spatially aligned with the backbone outputs C3–C5 and serve as the geometric priors for their respective levels.
For feature fusion, we adopt a top-down pathway and retain only P3–P5, which match a single, medium-scale bowl target. Let the backbone outputs at C3, C4, and C5 be
,
, and
, respectively. The top-level pyramid output
is first obtained by applying
and
convolutions to
to produce the base fused feature
, as defined in Eq. (10):
In these equations,
denotes the channel-compressed top-level feature, and
denotes the base fused feature at the top level.
and
represent
and
convolutions, respectively. For the intermediate level, taking the fourth backbone output
as an example, we first apply a
convolution to
. We then upsample the pyramid feature
by a factor of two so that its spatial resolution matches that of C4. The two features are added element-wise and passed through a
convolution to obtain the base fused feature for this level. This process is defined in Eqs. (11) and (12):
In these equations,
denotes the channel-compressed feature at the fourth level.
denotes a two-fold up sampling operation; here, it upsamples the pyramid output from the upper level—
in this example—by a factor of two, producing the upsampled feature
.
In these equations,
denotes the fused feature obtained by element-wise addition at the fourth level, and
denotes the base fused feature at this level. Subsequently, the GPM module applies a
convolution followed by a sigmoid activation to each level’s geometric prior feature
to produce a normalized geometric weight map. This weight map modulates the base feature
in a residual manner, yielding the geometry-modulated feature map
at level
. The process is defined in Eqs. (13) and (14):
In these equations,
denotes the pyramid level (here,
to
).
is the normalized geometric weight map at level
,
denotes the sigmoid function,
denotes the convolution operator,
is a
convolution kernel, and
is the geometric prior feature at level
.
In these equations, denotes the pyramid level (here, to ). is the feature map at level after geometric prior modulation. denotes the geometric prior modulation module. is the base feature at level , and is the geometric prior feature at the same level. denotes element-wise multiplication, and is the normalized geometric weight map for level .
Finally, we apply the CBAM module to
to adaptively reweight the features along both the channel and spatial dimensions, producing the geometry-aware pyramid output
. This process is given in Eq. (15):
In this equation, denotes the pyramid level (here, to ). is the geometry-aware pyramid output at level , and denotes the CBAM module.
2.4.4. Bowl Mask Segmentation Model Process
This subsection summarizes the overall workflow of the proposed bowl mask segmentation model based on an improved Mask R-CNN. First, we adopt the improved Mask R-CNN as the baseline framework. To meet the requirements of single-bowl images and high geometric precision, we optimize region proposal generation and introduce the GFPN structure to strengthen the model’s sensitivity to bowl contours. By incorporating the GPM module and the CBAM attention mechanism, the model can more accurately capture fine-grained features of the bowl rim, base, and sidewall. To further improve segmentation accuracy, GFPN enhances the bowl’s geometric boundaries across multiple scales and performs feature fusion on the outputs of the C3–C5 layers. With these optimizations, the model effectively reduces background noise interference and improves the accuracy of bowl contour segmentation.
Figure 8 illustrates the architecture of the improved bowl mask segmentation model.
2.4.5. Experimental Results and Analysis
In this experiment, we use 363 frontal-view images of bowls that were captured together with the corresponding top-down images, and expand the dataset to 1089 images through data augmentation. The dataset includes bowls of various sizes. All images were manually annotated using the LabelMe tool and converted to the standard COCO annotation format. Unlike the original COCO setting, this task does not involve bowl-type classification; therefore, all instances are labeled as a single class, while still retaining bounding boxes and segmentation masks. The hardware and software settings are the same as those in
Section 2.3. Each image has an original resolution of
pixels. The dataset was randomly split, with 80% of the data used for training and 20% used for testing.
In this study, we adopt a Mask R-CNN model pre-trained on the COCO dataset and fine-tune it by loading the corresponding pre-trained weights. To improve accuracy and efficiency, we adjust the RPN_ANCHOR_SCALES parameter to (16, 32, 64, 128, 256) to better accommodate feature extraction for smaller image sizes. The initial learning rate is set to 0.0005, with a momentum of 0.9, weight decay of 0.00005, a batch size of 8, and 100 training epochs to ensure stable optimization on the reduced image resolution. During training, we use a stepwise learning-rate decay schedule, reducing the learning rate by a factor of 0.1 every 15 epochs. To further improve training stability and convergence speed, we employ a staged training strategy. In the early stage, the first several layers of Res Net are frozen and only the subsequent layers are trained, allowing the network to focus on task-relevant features. As training progresses, more Res Net layers are gradually unfrozen, ultimately enabling end-to-end optimization of all layers.
For model evaluation, we adopt COCO-style segmentation metrics, including
(Intersection over Union),
(Mean Intersection over Union),
(Average Precision), and
(Mean Average Precision)[
38,
39]. Specifically, in this experiment,
is defined as the ratio between the area of intersection and the area of union of the predicted mask and the ground-truth mask, and
is the average
over all images. Since our task involves segmentation of only a single class,
is equivalent to
as used in multi-class settings. The calculation is given in Eq. (16):
To quantify the model’s boundary prediction accuracy, we evaluate it using the
with a 3-pixel tolerance. This metric assesses boundary performance by computing
, defined as the proportion of the predicted boundary that overlaps the ground-truth boundary, and
, defined as the proportion of the ground-truth boundary that overlaps the predicted boundary[
40]. The final
is the weighted harmonic mean of
and
, as given in Eq. (17):
In this experiment, we report and at two thresholds, 0.50 and 0.75, as well as the mean and averaged over these two thresholds. The corresponding metrics are , , , , , and , together with the .
To systematically evaluate the effects of different backbones and the proposed GFPN on bowl mask segmentation performance, we conduct a comparative study based on the improved Mask R-CNN baseline with different configurations. Three representative residual-network backbones are considered: Res Net-50, Res NeXt-50 (32×4d), and Res Net-C4. Here, Res Net-C4 serves as a shallower backbone without FPN, whereas the other two backbones are equipped with either the standard FPN or the proposed geometry-aware feature pyramid GFPN. All models are trained and evaluated on the same grayscale-processed frontal-view bowl dataset using the metrics described above.
Table 4 summarizes the comparative results of the bowl mask segmentation models.
Under the same experimental settings and hyperparameters, we select the best-performing configuration described above—using Res NeXt-50 (32×4d) as the backbone and GFPN as the FPN type—which we refer to as Ours, and compare it with other mask segmentation models, including U-Net, HTC, and B Mask R-CNN. All baseline models are evaluated using their original architectures without any module modifications. U-Net employs skip connections to fuse multi-scale features, balancing global semantics and local boundary details for pixel-level mask prediction[
41]. HTC alternately optimizes the detection and mask branches and progressively refines predictions through a multi-stage cascade, improving localization and segmentation quality[
42]. B Mask R-CNN is a boundary-enhanced variant of Mask R-CNN that introduces contour supervision and feature enhancement to strengthen edge representations and improve mask boundary accuracy and consistency[
43].
Table 5 reports the comparative results of these four models.
Figure 9 presents qualitative segmentation outputs for all models, along with the training and validation loss curves of our model and the
and
curves of the four models.
2.5. Estimation of the Volume of a Bowl
After obtaining the bowl mask, we use the segmentation model from the previous subsection and define 10 geometric keypoints on the mask: the outer-contour extremal points
and four uniformly sampled points along the right-side arc length,
. These keypoints are used to derive the rim diameter, base diameter, and bowl height, and to fit the inner-wall contour by incorporating the predicted wall thickness. Finally, the bowl volume is computed via axisymmetric integration[
44]. The detailed procedure is as follows:
1. Foreground pixel set and contour set: After bowl mask segmentation, let the binary mask be
. We first construct the foreground pixel set and the contour set on the mask, which are used for subsequent definition of geometric keypoints and fitting of the outer-wall curve, as given in Eq. (18):
In this equation, denotes the set of foreground pixels in the mask image. is the mask value at pixel , where the foreground is 1 and the background is 0, and and are the horizontal and vertical pixel coordinates in the image coordinate system. denotes the contour set extracted using OpenCV’s contour detection.
2. Six outer-contour extremal points
: To stably obtain the geometric boundaries of the bowl rim and base from the mask contour, we define four extremal points on
: the left and right upper-rim points
and
, and the left and right lower-base points
and
. The upper-rim points characterize the rim width, whereas the lower-base points characterize the base width and help determine the base position. Their definitions are given in Eqs. (19) and (20):
In this equation,
and
denote the minimum and maximum horizontal coordinates on the contour set
, respectively.
denotes the point within the specified domain that maximizes the objective function.
and
correspond to the left and right upper-rim points, respectively. Specifically,
and
are the highest points in the leftmost and rightmost contour columns, respectively.
In this equation, denotes the point within the specified domain that minimizes the objective function. and correspond to the left and right lower-base points, respectively. Specifically, and are the lowest points in the leftmost and rightmost contour columns, respectively.
Because the base primarily provides structural support and does not contribute to the effective holding volume, directly using
and
would lead to an overestimation of bowl height. Therefore, we analyze the variation of the mask’s horizontal width along the vertical direction to locate the shape transition between the base and the bowl body. The left and right boundaries at this transition are defined as
and
, which are used for estimating the effective height and fitting the sidewall. The definition is given in Eq. (21):
In this equation, denotes the foreground width of the mask at row , and and are the rightmost and leftmost horizontal coordinates of the foreground region at row , respectively.
In the base region,
varies only slightly, whereas once entering the bowl body sidewall,
increases markedly as
increases. We scan upward from the base and identify the first row at which the width change rate exceeds a threshold
, denoted as
. In our implementation,
is set to 5 px, and the left and right boundary points at this split row are defined as
, as given in Eq. (22):
In this equation, is the pixel -coordinate of the height where the base transitions to the bowl body. is the left boundary point at the start of the bowl body after removing the base, and is the corresponding right boundary point.
3. Pixel-scale geometric quantities: rim diameter, base diameter, and effective height. After obtaining the keypoints
, we can directly compute the bowl rim diameter, base diameter, and effective height in pixel units. The pixel-scale rim diameter is defined in Eq. (23):
In this equation,
denotes the rim diameter in pixels, and
and
are the horizontal coordinates of points
and
, respectively. The pixel-scale base diameter is defined in Eq. (24):
In this equation,
denotes the base diameter in pixels, and
and
are the horizontal coordinates of points
and
, respectively. The bowl’s effective height uses the midpoint of the rim as the upper reference point and the base–body split line defined by
and
as the lower reference, thereby reducing the influence of the base height. The pixel-scale effective height is defined in Eq. (25):
In this equation, denotes the effective bowl height in pixels, and , , , and are the vertical coordinates of , , , and , respectively.
4. Conversion from pixel units to physical scale: Using the pixel-to-world mapping obtained in
Section 2.1,
, we convert the pixel-scale measurements to real-world dimensions, including the rim diameter
, base diameter
, and bowl height
. The conversion is given in Eq. (26):
In this equation, , , and denote the rim diameter, base diameter, and bowl height in real-world units, respectively, while , , and denote the corresponding measurements in pixel units.
5. Sampling the right-side outer contour: Since a bowl can be approximated as an axisymmetric solid generated by rotation about its central axis, we model the bowl using the right-side outer contour. Specifically, we select the contour arc segment from
to
and denote the contour point sequence as
. Using arc-length parameterization, we uniformly sample four interior points
along this arc. Together with the endpoints
and
, these points form a representative set that characterizes the outer-wall shape, as defined in Eq. (27):
In this equation,
denotes the arc-length increment between adjacent contour points.
and
are the horizontal and vertical coordinates of the
-th point, and
and
are defined analogously. Here,
is the contour-point index ranging from 1 to
, where
is the total number of contour points.
represents the cumulative arc length from the starting point
to the
-th point.
denotes the arc-length increment between the
-th point and the
-th point, where
is the summation index ranging from 1 to
.
is the arc length at the starting point and is set to 0.
denotes the total arc length from the starting point
to the endpoint
, i.e., the full length of the selected contour segment. We then select four interior point locations by dividing the arc-length interval
into five equal parts, as defined in Eq. (28):
In this equation, denotes the arc-length position of the -th sampled point, where is the sampling index and takes values from 1 to 4. By performing linear interpolation on at , we obtain the sampled point , i.e., the coordinates of .
6. Outer-wall fitting and inner-wall correction: To convert the contour points into a continuous geometric representation, we determine the bowl’s central axis using the midpoint of the rim and project the right-side sampled points onto the height–radius plane. In physical units, we fit the outer-wall radius function
. We then combine this with the wall thickness
predicted by Bowl Thick Net to obtain the inner-wall radius function
, as defined in Eq. (29):
In this equation, denotes the outer-wall radius as a function of height , and denotes the inner-wall radius as a function of height . The coefficients , , , and are obtained via least-squares fitting.
7. Bowl capacity via axisymmetric integration: After obtaining the inner-wall profile
, we compute the bowl’s inner-cavity volume using the volume-of-revolution formula. The integration limits correspond to the effective height range, starting from the height of
at the bottom and ending at the rim at
. The formulation is given in Eq. (30):
In this equation, denotes the bowl capacity, and are the lower and upper bounds of the effective height in physical coordinates, and is the inner-wall radius function.
In summary, based on the frontal-view bowl images, we use the outputs of our improved mask segmentation model to extract key geometric parameters, including the rim diameter, base diameter, and bowl height. By further incorporating the wall thickness predicted by Bowl Thick Net, we construct axisymmetric contour curves for the inner and outer walls, derive an integral formulation for bowl capacity, and complete the bowl reconstruction model.
Figure 10 illustrates the bowl capacity estimation procedure and the corresponding results.
Following the above pipeline, we conduct a capacity validation experiment on eight bowls of different sizes. For each bowl, we first acquire a top-down image and a frontal-view image. The rim contour in the top-down image is used to predict wall thickness, while the mask and contour point set from the frontal-view image are used to obtain the bowl’s geometric parameters and contour curve. The outer-wall contour is then corrected to the inner-wall contour, yielding image-based geometric parameters, a reconstructed bowl model, and the predicted bowl capacity. As ground-truth references, we manually measure the rim diameter, base diameter, effective height, and actual holding volume of each bowl using tools such as vernier calipers and a graduated cylinder. Finally, we evaluate the accuracy of the proposed bowl model reconstruction and volume estimation method by comparing the predicted and measured geometric dimensions and capacities of the eight bowls on a per-bowl basis and conducting error analysis.
Table 6 presents the measured and predicted values of rim diameter, base diameter, effective height, and bowl capacity for each bowl.
As shown in
Table 6, the prediction error for the rim diameter of the bowl ranges from 2.19% to 4.67%, with an arithmetic mean error of 1.09%. For the base diameter of the bowl, the error ranges from 1.03% to 3.42%, with an arithmetic mean error of 1.14%. For the effective height of the bowl, the error ranges from −2.26% to 1.03%, with an arithmetic mean error of 0.40%. For the bowl volume, the error ranges from −7.29% to 3.41%, with an arithmetic mean error of −3.03%.