1. Introduction
Automated part feeding and orientation are essential in modern high-volume manufacturing and assembly lines. These processes directly affect throughput, cost, and product quality. Conveyor systems with passive geometric features, such as fences that act as geometric barriers along the conveyor, are a robust and cost-effective way to orient parts. They do not need complex robotics or sensors. However, designing these linear feeders is difficult. Engineers must choose the right type, number, and configuration of fences for each part shape to get the desired orientation. This is a major engineering challenge, particularly for complex 3D parts and intricate feeder designs. Traditional design relies on expert intuition and trial-and-error [
1], often leading to suboptimal designs and long development times. While physical simulations and algorithmic methods have been explored [
2,
3,
4], they often struggle with complex part–fence interactions, require significant computation for each new design, and may not predict feeder performance reliably before testing. This study focuses specifically on predicting orientation distributions for z-axis-symmetric 3D parts, extruded 2D shapes, interacting with linear feeders containing up to two straight or curved fences, laying the groundwork for more complex scenarios. Representative part–fence configurations are shown in
Figure 1.
Current solutions, such as dynamic simulations, geometric algorithms, and early machine learning methods, have improved the field. However, there is still a major gap. To the authors’ knowledge, there are no widely adopted automated tools that accurately predict the full probability distribution of final part orientations, that is, the likelihood of a part ending in any possible orientation after passing through the feeder, for a given part and feeder design across diverse geometries and fence configurations. Predicting this full orientation distribution, rather than just a single likely outcome, is crucial for robust design and performance assessment. Furthermore, the high computational cost of simulation hinders exploration, and there are few methods to optimize this data-generation process. Researchers also need methods that use these predictions to help automate the feeder design process, not just analyze it. Most simulation tools require many runs for each design, and current machine learning approaches do not generalize well to new part shapes or give a strong measure of part–feeder compatibility.
This research proposes a new approach to fill this gap. We combine physics-based simulation with deep learning, using Variational Autoencoders, or VAEs, and regression models. These models learn the mapping between part geometry, fence configuration, and the resulting orientation distribution, represented as a discrete 360-bin Probability Mass Function, PMF, and its circular CDF. A key aspect is exploring the VAE not only for learning joint representations of parts, feeders, and orientation PMFs and CDFs, but also as a first step toward more efficient simulation pipelines. Specifically, we investigate training on less converged simulation data, 5–100% of 1000 iterations in 5% increments, and report reconstruction performance as a function of convergence level, alongside a cross-convergence evaluation on the 100% holdout. This test remains limited to reconstruction on the same configuration set. The PMF assigns probability to each orientation bin, while the CDF summarizes cumulative probability around the circle. This learned representation, together with direct regression models, is used to predict feeder performance and support a more informed design process, serving as a step toward automation.
The research objectives are:
- 1.
Generate a comprehensive dataset of part-orientation distributions using physics-based simulation in CoppeliaSim for z-axis-symmetric 3D parts interacting with linear feeders.
- 2.
Develop and evaluate deep learning models, regression networks, and VAE, to predict orientation distributions with target accuracy on unseen configurations.
- 3.
Investigate the VAE’s ability to learn latent representations from simulations at different iteration counts, 5–100%, quantifying within-level and cross-convergence reconstruction performance to assess whether partial simulations can support cost reduction.
- 4.
Evaluate a delta-to-full correction model that predicts the update from partial to fully converged PMFs using multiple checkpoints, under part-level and configuration-level splits.
- 5.
Assess model generalization to unseen feeder configurations and new part geometries, quantifying performance degradation and identifying limiting factors.
This research uses a new simulation-driven deep learning approach to solve these challenges. The following sections explain the methods used, including the simulation setup; data representation with Discrete Fourier Transform (DFT) for parts, one-hot encoding for feeders, and PMF/CDF for distributions; model architectures with fully connected networks and VAE; and evaluation strategies. We compare different modeling approaches and test their performance on various datasets, including new parts and robustness to changes. The discussion explains the results, points out limitations, and suggests future research for real-world use. The paper is organized as follows:
Section 2 reviews related work.
Section 3 explains data generation, representation, and model development.
Section 4 presents model results.
Section 5 discusses the results, limitations, and future work.
Section 6 summarizes the main contributions and findings.
Key contributions.
Dataset: A simulation-driven dataset of 1,236 part–feeder configurations, split into Main, Test, Fences, and Parts datasets, with 360-bin orientation PMFs and CDFs for extruded 2D parts and up to two fences.
Regression models: Models that predict full orientation distributions, achieving on circular CDFs of 0.97 on held-out test configurations and 0.89 on unseen fence combinations for known parts, with a marked drop on unseen parts to 0.75.
VAE convergence study: A VAE reconstruction study across partial-iteration datasets (5–100% in 5% increments) showing high within-level reconstruction () but weak cross-convergence to 100% labels (overall CDF = 0.01 at 5%, 0.32 at 50%, and 0.87 at 75%), consistent with partial-iteration PMFs remaining far from fully converged distributions.
Delta-to-full correction: A correction model that predicts the update from partial PMFs to the 100% PMF, improving low-iteration performance (e.g., at 5% for unseen parts vs. a negative baseline), with a full 5% sweep reported in Supplementary Section S11.
Scope: We consider extruded 2D, z-axis-symmetric parts and up to two straight/curved fences; broader 3D parts and more complex feeders are discussed in
Section 5.
1.1. Problem Formulation
We consider a supervised prediction task from part and feeder descriptors to an orientation distribution. For each configuration, the input vector is
where
encodes the part geometry using DFT coefficients and
encodes the fence types and angles. The target output is the discrete orientation distribution obtained from simulation:
where the PMF assigns probability to each
bin and the CDF is its circular cumulative version. Regression models learn a mapping
, while the VAE learns a joint latent representation to reconstruct
and evaluate reconstruction fidelity.
4. Results
4.1. Experimental Results Summary
Results are organized by task: training fit, generalization to held-out configurations and unseen fence combinations, generalization to unseen parts, and VAE reconstruction performance. Metrics are reported as (mean ± SD) on circular CDFs unless stated otherwise.
4.2. Regression Model Performance on Training Data
We trained all eight regression models on 85% of the generated dataset (~890 samples), optimizing hyperparameters as detailed previously.
Table 5 summarizes the training performance, measured by
and
(mean ± SD) between predicted and true distributions (PMF/PCA). Training loss curves for Model 3 are provided in Supplementary Figure S4. Throughout this subsection, “PMF” refers to the 360-bin histogram (probability per
bin), so per-bin values are unitless and sum to 1.
As shown in
Table 5, most models achieved high
values on the training data (0.97–0.98), while the PMF-PCA models (1–2) were lower (0.86–0.90), indicating reduced fit for the PCA-based label representation. Despite similar quantitative performance among the remaining models, qualitative differences in the predicted distributions were observed. After applying the output post-processing described in
Section 3.4.1, the models exhibited the following patterns:
Models 1 and 2 (PMF-PCA): Showed occasional local oscillations and spiky bins after reconstruction; Model 2 (branched architecture) exhibited fewer and smaller artifacts than Model 1 (standard architecture).
Models 3 and 4 (PMF-Full): Produced smooth predictions with no per-bin values exceeding 1 (due to the Softmax output). Because we treat the 360-bin histogram as a PMF, any per-bin probability >1 is invalid; Softmax heads prevent this by construction. However, predictions were sometimes overly smooth compared to the true distribution, particularly in regions of sharp probability increases.
Models 5 and 6 (CDF-PCA): When converted back to PMFs for visualization, these models exhibited artifacts. Model 6 (branched) had larger but less frequent artifacts than Model 5 (standard).
Models 7 and 8 (CDF-Full): Converted PMFs showed fewer artifacts than the PCA-based CDF models. Model 7 (standard) had more residual oscillations when converted back to PMFs than Model 8 (branched).
Representative training predictions for Model 3 are shown in Supplementary Figure S5. These initial observations suggest that while all models learn the training data well quantitatively, the choice of label representation (Full vs. PCA) and output activation (Softmax vs. Sigmoid/Linear) significantly impacts the qualitative validity (e.g., monotonicity) of the predictions, even on familiar data.
4.3. Generalization to New Fence Angles (Test Dataset)
Protocol P1: The Test dataset evaluated generalization to minor variations using known part–feeder configurations, but with fence angles not seen during training. This aligns with P1; see
Section 3.5.1.
Table 6 summarizes the performance of the models on this dataset with
and
.
We evaluated fence generalization on known parts by holding out fence angles and configurations not seen during training. Across the test dataset, the regression models achieved values ranging from 0.85 to 0.98, supporting the claim that the models learned fence effects independently of geometry when the geometry was represented in the training distribution.
Performance remained high across most models, with values ranging from 0.85 to 0.98 and only modest drops relative to training. The standard deviations increased slightly. Qualitative behaviors observed in training largely persisted. Models 3 and 4 (PMF-Full) showed slightly more pronounced smoothness, while Models 5 and 6 (CDF-PCA) showed some artifacts when converted back to PMFs. The minimal performance drop indicates that the models generalize well to small perturbations in continuous parameters, such as the fence angle. These angle changes correspond to relatively small shifts in the 53-dimensional input feature space; we did not quantify this sensitivity. Per-part breakdowns for the harder generalization splits are reported in Supplementary Table S2 to diagnose failure modes beyond the aggregate means.
Figure 12 provides representative predictions.
4.4. Generalization to New Feeder Configurations (Fences Dataset)
Protocol P1: The Fences dataset tested generalization to novel combinations of known parts and known feeder types or locations. See
Section 4.3 for protocol and aggregated metrics; here we evaluate unseen fence
combinations rather than just angles.
Table 7 summarizes model performance on this dataset with
and
, showing a clear drop relative to the test-angle split.
Representative predictions for the Fences dataset are provided in Supplementary Figure S6. Performance decreased compared to the Test dataset but remained relatively strong ( generally –). Standard deviations increased further. Qualitative issues became slightly more apparent: Model 2 (PMF-PCA, branched) showed more oscillations, approaching those of Model 1. Models 6 and 8 (CDF-PCA branched and CDF-Full branched) exhibited larger artifacts after conversion to PMFs. This indicates that predicting the outcome of entirely new part–feeder combinations, even with familiar components, is more challenging than interpolating angles. The models still capture the general distribution shape reasonably well. Per-fence-type breakdowns show that, for Model 3, Curved-1 + Curved-2 configurations are the most difficult (, ), while Straight-1 + Straight-2 and Curved-2-only configurations are easiest (, –).
4.5. Generalization to New Part Geometries (Parts Dataset)
Protocol P3: This dataset represents the most challenging generalization task, using entirely new part geometries not seen during training, combined with known fence configurations; see
Section 3.5.1. Only three parts were held out,
Table 4, so this provides a limited probe of zero-shot generalization.
Table 8 summarizes performance with
and
.
For the Parts dataset, the full-output and CDF-based models showed a clear drop in performance (
roughly 0.72–0.75) with substantially increased standard deviations. Model 3 (PMF-Full, standard architecture) remained among the better full-output models (
), but the overall accuracy is considerably lower than for other generalization tasks. The PMF-PCA models (1–2) achieved higher
on CDFs (∼0.94). However, their reconstructions still exhibited PCA-related artifacts (oscillations/smoothing) when converted back to PMFs, so the elevated CDF
should be interpreted cautiously. As a sanity check, a mean-distribution baseline (using the training-set mean PMF and evaluating on Parts CDFs) yields
, indicating that the Parts split has relatively low variance but that the PCA models still exceed a trivial predictor. Qualitative behaviors persisted, and the predicted distributions often failed to accurately capture the shape and location of probability mass in the true distributions (
Figure 13). Given the small number of unseen parts (3), these results should be interpreted as preliminary rather than definitive evidence about generalization to arbitrary new shapes. Per-part breakdowns (Supplementary Table S2) show that Model 3 performs worst on part 9 (
,
) and best on part 21 (
), highlighting part-specific failure modes within the limited unseen set.
Diagnosis: This poor generalization to new parts reflects limitations in the chosen part representation (DFT) and the models’ ability to extrapolate beyond the geometric variations present in the training data. The 45 DFT parameters, while compact, do not capture enough of the salient features relevant to dynamic interaction, and the training set lacks sufficient geometric diversity. The models overfit to the specific part geometries seen during training, struggling when presented with substantially different shapes. Expanding the Parts dataset with more diverse unseen geometries is necessary before drawing stronger generalization conclusions.
4.6. Geometric Diversity and Noise Robustness
We analyzed geometric diversity and robustness to feature noise to understand generalization limits better. The full analysis, including descriptive statistics, correlation plots, and noise-injection results, is provided in Supplementary Sections S7–S8 (Figures S7–S8, Tables S1–S3). In brief, the training set underrepresents highly concave and high-aspect-ratio geometries, and the observed correlations between geometric metrics and on the Parts dataset are weak: compactness (), convexity (), and aspect ratio () across part–feeder pairs. Model 3 remains relatively robust to moderate feature noise.
4.7. VAE Performance
While the regression models demonstrated strong performance for known configurations and robustness to noise, they exhibited significant limitations in generalizing to new part geometries (
Section 4.5). This suggests fundamental constraints in either the part representation or the models’ ability to learn generalizable physical principles from the available data. Additionally, the data generation process—requiring 1000 simulation iterations per configuration—presents a computational bottleneck for scaling to more diverse parts and configurations. To address these limitations, we explored a complementary approach using a Variational Autoencoder (VAE), which offers potential advantages through its ability to: (1) learn a compact latent representation capturing the joint distribution of parts, feeders, and orientation PMFs/CDFs; (2) evaluate reconstruction when trained on less converged simulation data, including a cross-convergence test against the 100% holdout, as a first step toward potential cost reduction; and (3) provide a mechanism for validating configurations through reconstruction quality. The following results evaluate whether the VAE can overcome the generalization challenges faced by the regression models while offering these additional capabilities.
We trained the VAE on the Main dataset only (
configurations with part+feeder features and full CDF labels), optimizing hyperparameters as detailed in
Section 3.4.2. The Fences and Parts datasets were held out entirely from VAE training and used solely for evaluation in
Table 9.
Important: this VAE is an autoencoder that reconstructs its inputs (part+feeder features and the CDF) and is not a conditional model that predicts a distribution from geometry alone; thus
Table 9 reports reconstruction quality, not predictive performance from geometry.
4.7.1. VAE Training and Reconstruction Performance
The VAE training converged well, with loss components balancing reconstruction and KL divergence (Supplementary Figure S9). Reconstruction accuracy (
) was evaluated on the held-out Fences and Parts datasets (
Table 9).
The VAE achieved reasonable reconstruction performance on the Fences dataset (–), but, similar to the regression models, struggled on the Parts dataset (–). This reinforces the difficulty of generalizing to unseen part geometries.
4.7.2. Performance with Partial Simulation Data
We evaluated VAE reconstruction across different simulation iteration counts to assess whether training on partially iterated data could reduce computational cost. This setup corresponds to P2; see
Section 3.5.1. We trained separate VAE models from scratch for each of 20 iteration levels (5% to 100% in 5% increments) using datasets rebuilt from the same configurations but with CDFs computed from the corresponding iteration count. The partial-iteration CDFs replaced the full-iteration labels for that run with no mixing or augmentation, so dataset size remained fixed across levels at
configurations (1,048 Main plus 110 Parts, excluding 78 Fences because raw per-iteration traces are unavailable). Note that this convergence study is not a generalization test: it uses all configurations for which per-iteration traces exist, including Parts, to assess reconstruction quality as a function of iteration count rather than zero-shot prediction on unseen geometries. For fair comparison, we created one 80/20 split with a fixed seed, trained each VAE on the shared 80% subset, and evaluated on the shared 20% holdout (
).
Within-level vs. cross-convergence evaluation. To directly test partial-to-full reconstruction, we trained separate VAE models at 20 iteration levels (5% to 100% in 5% increments) and evaluated each on the shared holdout using the 100% CDFs as targets. Thus, the input is the partial PMF and the target is the 100% CDF.
Figure 14 compares cross-convergence performance (red) against within-level reconstruction (green). Performance improves monotonically with iteration count:
remains below 0.10 up to 30%, rises through 0.32 at 50%, and reaches 0.87 at 75% before converging to 0.98 at 100%. The contrast between the flat within-level curve and the rising cross-convergence curve visually confirms that the extrapolation task is inherently difficult: the VAE architecture works well when labels match, but cannot fully converge on CDFs from partial-iteration inputs until approximately 70% of iterations are used.
4.7.3. Delta-to-Full Correction Model
To directly improve partial-to-full prediction, we trained a feedforward model to predict the correction
PMF such that
. The input combined part+feeder features with PMFs from up to three checkpoints (
in percentage-point units; zero-padded when
), their differences, and summary stability features (JS,
,
, entropy, peak mass). We evaluated a 5% sweep from 5% to 100% using two split modes: part-level GroupKFold (unseen parts) and random configuration splits (seen parts with new fence/angle combinations).
Table 10 reports representative levels; the full 5% sweep and curves are in Supplementary Section S11.
At low iteration counts, the delta model substantially improves over the baseline that uses the partial PMF as the final answer. For example, at 5% (95% cost saved), the part-level split achieves versus a baseline of , and the config-level split achieves versus . This indicates that predicting the correction is an effective strategy for cost reduction, though accuracy remains below full-iteration performance and should be interpreted relative to the application tolerance.
5. Discussion
We investigated the feasibility of using simulation-driven deep learning models, specifically regression networks and a Variational Autoencoder, VAE, to predict final part orientation distributions in linear conveyor feeders. The results demonstrate promising capabilities but also highlight significant challenges, particularly concerning generalization to novel part geometries.
5.1. Interpretation of Regression Model Performance
The eight regression models explored various architectural choices and label representations (see
Table 2). Model 3, a standard architecture that predicts full PMF with a Softmax output, emerged as the most consistently well-behaved and quantitatively accurate model across the evaluation datasets, particularly in terms of generalization and robustness. Its
values were high: training
in
Table 5; test
in
Table 6; fences
in
Table 7. Models 3 and 4 employed a composite loss function incorporating MSE on circular CDFs, defined as cumulative sums, KL divergence between PMFs, and the Kolmogorov–Smirnov statistic on circular CDFs, with
, ensuring both accurate probabilistic matching and proper cumulative distribution behavior. The use of a Softmax output layer inherently enforces non-negativity and the sum-to-one property for the predicted PMF. While Models 3 and 4, the PMF-Full variants, sometimes produced overly smooth predictions in Supplementary Figure S5 and
Figure 12 and
Figure 13, this was preferable to the invalid distributions generated by other models.
The branched architectures, Models 2, 4, 6, 8, did not consistently outperform their standard counterparts, Models 1, 3, 5, 7, suggesting that separating part and feeder feature processing offered no significant advantage for this task. Similarly, using PCA representations, Models 1, 2, 5–8, did not improve performance and often introduced artifacts like oscillations or spiky bins upon reconstruction, indicating that learning the full 360-dimensional distribution, despite its higher dimensionality, was more effective, especially when coupled with appropriate output activations such as Softmax for PMFs.
The models demonstrated strong generalization to variations in continuous parameters such as fence angles on the Test dataset, with , and reasonable generalization to new combinations of known parts and feeders on the Fences dataset, . This aligns with expectations, as these tasks involve interpolation or recombination within the learned feature space.
5.2. Addressing the Generalization Challenge
The most critical finding is the significant drop in performance when generalizing to entirely new part geometries on the Parts dataset,
for Model 3 and 0.72–0.75 for other full-output and CDF models, as reported in
Table 8. The PMF-PCA models yield higher CDF
but exhibit reconstruction artifacts, so the underlying generalization challenge remains. This suggests that the chosen part representation of 15 DFT coefficients, along with the diversity of geometries in the training set, was insufficient. While DFT captures basic shape information, it might miss subtle geometric features crucial for predicting complex dynamic interactions with fences. The models likely learned correlations specific to the training parts rather than generalizable physics principles applicable to any shape. This contrasts with some algorithmic approaches [
4] that aim for completeness under idealized conditions but struggle with real-world physics. In contrast, our data-driven approach captures physics implicitly via simulation but struggles with geometric extrapolation. This finding echoes challenges seen in other domains where ML models fail to generalize beyond the training distribution, particularly when relying on potentially incomplete feature representations.
The robustness analysis in Supplementary Section S8 showed that Model 3 was relatively insensitive to noise added to the part features, with
even with
in Table S3. This indicates the model is not overly sensitive to minor geometric imperfections, a positive sign for practical applicability. Crucially, this robustness indicates that the reduced performance on unseen parts (
Table 8) reflects fundamental limits in shape extrapolation rather than model brittleness—the model has learned genuine part–feeder relationships but cannot extrapolate to geometries outside the training distribution.
We emphasize that robust generalization to arbitrary unseen part geometries is not claimed as a contribution of this work. The Parts dataset (3 unique parts, 110 configurations) serves as a diagnostic probe to assess whether the learned relationships extrapolate beyond the training distribution. The substantial performance reduction (from
to
) confirms that they do not—a finding that motivates the representation and dataset improvements outlined in
Section 5.6. Expanding the Parts dataset would better quantify the generalization gap but would not resolve it without addressing the underlying representation limitations identified here.
Scope. While promising for known geometries and moderate variations, the present models do not yet generalize reliably to novel parts; see
Table 8. For manufacturing lines with a fixed catalog of parts, the method provides fast distribution prediction across fence parameters and combinations, enabling design-space exploration and QA; the remaining challenge is extrapolation to truly novel geometries.
5.3. VAE Insights and Potential
The VAE demonstrated its ability to learn a joint representation of parts, feeders, and CDFs, achieving reasonable reconstruction accuracy on known configurations in the Fences dataset, with
values of 0.90–0.96 in
Table 9. These results are reconstruction metrics from autoencoding—the VAE receives the target CDF as part of its input and is not asked to predict the distribution from geometry alone. Its struggle with new parts in the Parts dataset,
–
, mirrored the regression models, confirming the generalization difficulty.
Convergence rate and iteration-count experiments. The iteration-count results should be interpreted in light of the convergence distribution documented in
Section 3.1.3 and Supplementary Figure S10. At 5% iterations, none of the 1,158 configurations meet JS
, and at 50%, only one does. This explains why within-level reconstruction remains high (green curve in
Figure 14) while cross-convergence performance is low (red curve; overall CDF
= 0.01 at 5%, 0.32 at 50%, and 0.87 at 75%).
Importantly, these results do not demonstrate that partial simulations can, in general, replace full simulations to reduce computational cost. Cross-convergence performance is low across the board (overall CDF at 50%), indicating that partial-iteration labels remain far from fully converged labels for most configurations. The VAE’s reconstruction quality could still serve as a configuration assessment metric, evaluating the plausibility of a predicted or designed part–feeder–distribution triplet based on its proximity to the learned data manifold.
Delta-to-full correction for cost reduction. The delta-to-full model targets cost reduction by estimating the correction from a partial PMF to the fully converged PMF using multiple checkpoints. This yields large gains at low iteration counts: at 5% iterations,
improves from negative baseline values to about 0.82 (part split) and 0.83 (config split), while at 50% it improves from 0.31 to about 0.86 (
Table 10, Supplementary Section S11). These results indicate that learning the correction is an effective strategy for cost reduction. However, accuracy at very low iteration counts remains below full-iteration performance and should be judged against the application’s tolerance.
The VAE remains useful for denoising, anomaly detection, and comparing configurations, even if it cannot extrapolate from partial to full iterations. Combining partial-iteration training with active learning or Bayesian optimization to select which configurations to simulate is a potential avenue for future research.
5.4. Comparison with Literature
This work advances beyond previous simulation-based methods [
2,
3,
18] by using deep learning to create predictive models rather than just analyzing individual simulations. Unlike purely algorithmic approaches [
4,
14,
15,
16,
17] that often rely on simplified physics and typically aim for a single target orientation, such as Wiegley et al. [
4], our method implicitly captures complex dynamics through simulation and explicitly predicts the full probability distribution of final orientations. This provides a richer understanding of feeder performance and robustness compared to methods focused solely on achieving a deterministic outcome. Compared to prior ML applications in part feeding—orientation recognition [
22,
23] and RL-based trap configuration for vibratory bowl feeders [
24]—this study tackles the prediction of the full orientation distribution, a more complex regression problem essential for design evaluation. However, the generalization limitations observed highlight that our data-driven approach does not yet match the theoretical completeness of some algorithms [
4] for the geometries they cover, nor the flexibility of robotic systems [
11,
12]. The partial-iteration VAE results show strong reconstruction within matched convergence levels but low cross-convergence performance across the dataset, so they do not establish that fewer iterations can replace fully converged labels or reduce simulation cost without loss.
The iteration-count findings also highlight a methodological consideration for simulation-driven ML: characterizing convergence heterogeneity in
Section 3.1.3 is essential before claiming computational savings from partial simulations.
5.5. Limitations and Simulation-Reality Gap
Several limitations must be acknowledged.
Physics model limitations. A primary limitation is the study’s exclusive focus on z-axis-symmetric 3D parts that are extruded 2D shapes and simplified linear feeders containing only up to two fences that are straight or curved. This scope neglects the significant complexities associated with fully 3D parts, including out-of-plane rotations and non-uniform cross-sections, as well as more intricate feeder designs commonly found in real-world applications, thereby limiting the direct generalizability of the current findings.
DFT imposes a global, frequency-domain representation that limits the capture of local contact geometry. Its partial rotation normalization (magnitude invariance plus canonical alignment) is beneficial for shape recognition but can undermine predictions that depend on absolute orientation and directional contacts. The DFT representation emphasizes global features (low-frequency coefficients) over local geometric details (high-frequency coefficients). Our selection of only 15 coefficients further prioritizes global shape characteristics at the expense of fine details that influence part–fence interactions, such as small protrusions, corners, or subtle curvature changes. DFT’s spectral nature also makes it difficult to spatially localize specific geometric features, obscuring the precise contact points and interaction mechanics between parts and fences.
Representation limits are most visible in the new-shape split. The generalization gap for unseen parts aligns with the fact that the training set underrepresents the geometric features most likely to drive new contact behaviors.
Alternative representations can capture more spatial detail. Point clouds preserve explicit geometry, and graph-based representations encode local connectivity. These choices introduce their own challenges, particularly regarding rotation invariance and computational complexity.
Dataset limitations. The dataset size (1,236 samples total: 1,048 main + 78 fences + 110 parts) was likely insufficient in both quantity and geometric diversity to enable robust generalization, especially given the high-dimensional nature of the part–feeder distribution relationship.
Contact/friction limitations. Bullet uses a simplified Coulomb friction model with constant static and kinetic coefficients, which omits several real-world effects: (1) stiction, where initial resistance to motion exceeds kinetic friction; (2) velocity-dependent friction; (3) material-dependent hysteresis and compliance; and (4) anisotropic friction properties [
29,
31]. These omissions limit fidelity for contacts sensitive to microslip and surface conditions.
Additional physics simplifications compound this gap. All parts are modeled as rigid bodies, omitting elastic deformation that can affect part–fence interactions, especially for complex geometries or compliant materials. The simulations also omit multi-part interactions, idealize conveyor belt properties (texture, compliance, vibration), and ignore environmental factors such as humidity and electrostatic effects.
Reality gap implications. The simulation simplifications limit deployment confidence. Models can underperform on physical systems because unmodeled dynamics shift the true distributions, and certain geometries are especially sensitive to these effects, which aligns with the observed generalization failures on new shapes.
The lack of physical validation is the dominant gap. Without experiments, it is not possible to separate modeling limitations from simulation-physics mismatch or to quantify the role of parameters such as friction and compliance. This motivates the targeted experimental validation steps outlined below.
5.6. Future Work
Priority one is expanding the dataset’s geometric coverage. The next step is to add highly concave and high-aspect-ratio parts and to sample underrepresented regions of the shape metric space identified in
Section 4.6. This directly targets the failure modes observed on the Parts split.
The current correlations are based on only three unseen parts. A larger and more diverse Parts dataset is required to validate and refine the observed relationships between geometric metrics and prediction error, and to update the diversity metrics with stronger statistical support.
Active learning offers a principled path to dataset expansion. Model uncertainty can guide the selection of new simulation candidates, and acquisition functions can target parts with maximum expected information gain. This approach is more efficient than random sampling and should improve generalization with fewer additional simulations.
Generative sampling can complement active learning. The VAE’s latent space can be used to propose novel parts that fill geometric gaps, creating a closed-loop enrichment process.
Representation upgrades are central to improving generalization. DFT offers compactness but misses local features and topological detail that drive contact dynamics, which aligns with the observed failures on unseen shapes.
Geometric deep learning provides a viable alternative. Point clouds or mesh-based GNNs can learn localized features directly, and spectral descriptors (e.g., Laplace–Beltrami) offer intrinsic invariances that may improve extrapolation. These options should be tested to determine whether they yield more generalizable relationships between part geometry and orientation dynamics.
A primary direction is to overcome the significant limitation of the current 2D focus by extending the methodology to handle 3D part geometries and their complex interactions, including out-of-plane rotations. As the DFT representation used here proved insufficient even for 2D generalization, this necessitates exploring more powerful 3D shape representations capable of capturing dynamically relevant features. Promising alternatives include point cloud-based methods like PointNet++, graph neural networks operating directly on mesh data, or voxel-based representations combined with 3D convolutional neural networks.
Improving simulation fidelity is another parallel track. Incorporating more accurate friction models, part flexibility, and multi-part interactions will increase realism and reduce the simulation-reality gap. Physics-informed losses or PINN-style constraints can further regularize predictions toward physically plausible behavior.
A sensitivity study would quantify the robustness of the reported distributions to variations in the physics parameters. Such a study could vary friction coefficients by (parts, belt, fences), restitution e from 0.0 to 0.1, belt speed by , and timestep/substeps by and on a fixed subset of configurations (e.g., 30 randomly selected part–fence pairs). For each sweep, recomputing the 360-bin PMFs and reporting distributional changes (JS divergence or Wasserstein distance) relative to the baseline would identify which parameters materially affect the distribution and provide error bars on simulation-driven results without requiring full experimental calibration.
Physical validation is essential. A dedicated testbed and targeted experiments are required to quantify the simulation-reality gap and to confirm predictive accuracy on real systems.
6. Conclusions
This research investigated the application of deep learning models trained on physics-based simulation data to predict final part orientation distributions in linear conveyor feeders. We developed and evaluated multiple regression network architectures and a Variational Autoencoder (VAE), assessing their accuracy, generalization capabilities, and robustness.
The key findings show that a standard fully connected regression network predicting the full PMF using a Softmax output (Model 3) provided the most reliable and accurate predictions, achieving high values for known parts with varying fence angles () and new combinations of known parts and feeders (). This model demonstrated robustness to noise in part geometry, indicating practical applicability within the tested noise range. However, all models, including the VAE, exhibited significant limitations in generalizing to entirely new part geometries ( dropping to 0.75 for Model 3), due to limitations in the DFT part representation and insufficient geometric diversity in the training data.
From a practical design standpoint, a practitioner can use the current pipeline to screen and rank candidate feeder configurations for known parts before committing to physical prototypes. Specifically, for a fixed part library, the regression model can rapidly predict orientation PMFs/CDFs across a grid of fence angles and types, enabling selection of configurations that maximize desired orientation probability or minimize multi-modal outcomes. For the VAE, within-level reconstruction remains high, but cross-convergence results are low and do not demonstrate that fewer iterations can replace fully converged labels for predictive use; see
Section 3.1.3 and
Figure 14. However, the delta-to-full correction model showed substantial improvement over the baseline at low iteration counts, achieving
at 5% iterations versus a baseline of
, indicating that predicting the correction from partial to full PMFs is an effective strategy for reducing simulation cost; see
Section 4.7.3 and
Table 10.
For deployment, the next step is controlled physical validation to quantify the simulation–reality gap and calibrate the model. A concrete plan is to build a testbed linear feeder with interchangeable straight/curved fences, select a small but diverse subset of parts (including high-aspect-ratio and concave shapes), and collect empirical orientation distributions across a designed set of fence angles. These measurements should be used to (i) evaluate predictive accuracy on real data using the same CDF-based metrics, (ii) tune friction/restitution parameters and contact models to reduce systematic error, and (iii) test sensitivity to unmodeled effects (surface finish, belt compliance, part-to-part variability). Once validated, the model can be updated via domain adaptation or retraining to support deployment on production feeders.
In conclusion, this study demonstrates the feasibility of using simulation-driven deep learning to predict part orientation distributions in linear feeders for specific configurations, offering a useful tool for analysis within the scope tested in
Section 4.3Section 4.4. Given the limited generalization to unseen geometries and the absence of experimental verification, we position this method as an analysis aid rather than an automated design tool at this stage. While the VAE-based approach offers a novel way to model distributions and explore the design space, current evidence does not establish reduced-iteration simulation as a general cost-saving strategy. The significant limitations in generalizing to novel parts and the lack of physical verification currently restrict direct application to automated design. This work provides an initial foundation, but substantial future development, including addressing these key challenges through advanced representations, larger and more diverse datasets, and experimental verification, is required before this approach can reliably automate design processes for diverse industrial part feeding systems.
Figure 1.
Representative 2-D part silhouettes, top row, and their corresponding feeder configurations, bottom row. From left to right: panel a shows a simple convex polygon with a single straight fence; panel b shows an I-shaped profile with dual curved fences; panel c shows a highly asymmetric free-form part interacting with a mixed straight–curved layout. This visual taxonomy motivates the geometric diversity tackled in the present study.
Figure 1.
Representative 2-D part silhouettes, top row, and their corresponding feeder configurations, bottom row. From left to right: panel a shows a simple convex polygon with a single straight fence; panel b shows an I-shaped profile with dual curved fences; panel c shows a highly asymmetric free-form part interacting with a mixed straight–curved layout. This visual taxonomy motivates the geometric diversity tackled in the present study.
Figure 2.
Block diagram of the overall workflow illustrating the pipeline from STL import and physics simulation in CoppeliaSim to data representation using DFT, one-hot encoding, and PMF/CDF labels.
Figure 2.
Block diagram of the overall workflow illustrating the pipeline from STL import and physics simulation in CoppeliaSim to data representation using DFT, one-hot encoding, and PMF/CDF labels.
Figure 3.
Simulation overview. Panel a shows the CoppeliaSim simulation environment with a part on the linear feeder belt approaching a feeder configuration. Panel b shows time-stamped snapshots from 0 to 1800 ms of a typical simulation run, showing the part’s trajectory, fence contact, and final rest configuration. The sequence highlights stochastic elements such as initial orientation randomization and micro-impacts captured by Bullet and subsequently encoded in the orientation PMF.
Figure 3.
Simulation overview. Panel a shows the CoppeliaSim simulation environment with a part on the linear feeder belt approaching a feeder configuration. Panel b shows time-stamped snapshots from 0 to 1800 ms of a typical simulation run, showing the part’s trajectory, fence contact, and final rest configuration. The sequence highlights stochastic elements such as initial orientation randomization and micro-impacts captured by Bullet and subsequently encoded in the orientation PMF.
Figure 4.
Jensen–Shannon, JS, divergence between orientation distributions generated with N iterations and the full 1000-iteration reference. The curve shows mean ± SD across 10 randomly selected part–feeder configurations; see Equation
3. Note that this averaged view hides the broad distribution of JS values across all configurations.
Figure 4.
Jensen–Shannon, JS, divergence between orientation distributions generated with N iterations and the full 1000-iteration reference. The curve shows mean ± SD across 10 randomly selected part–feeder configurations; see Equation
3. Note that this averaged view hides the broad distribution of JS values across all configurations.
Figure 5.
Distribution of Jensen–Shannon (JS) divergence between the 100-iteration and 1000-iteration orientation PMFs across configurations. No configurations satisfy JS at 100 iterations (median JS ).
Figure 5.
Distribution of Jensen–Shannon (JS) divergence between the 100-iteration and 1000-iteration orientation PMFs across configurations. No configurations satisfy JS at 100 iterations (median JS ).
Figure 6.
DFT-based part representation. Panel a shows a conceptual illustration of representing a 2D part shape, blue outline, using a finite number of Discrete Fourier Transform coefficients, epicycles. Panel b shows examples of DFT reconstruction using 15 coefficients for diverse part shapes. Original outlines, blue, are closely approximated by the reconstructed shapes, solid orange, demonstrating the representation’s fidelity. Axes are in normalized part-coordinate units, unitless.
Figure 6.
DFT-based part representation. Panel a shows a conceptual illustration of representing a 2D part shape, blue outline, using a finite number of Discrete Fourier Transform coefficients, epicycles. Panel b shows examples of DFT reconstruction using 15 coefficients for diverse part shapes. Original outlines, blue, are closely approximated by the reconstructed shapes, solid orange, demonstrating the representation’s fidelity. Axes are in normalized part-coordinate units, unitless.
Figure 7.
Fence locations and encoding: Location 1 at m and Location 2 at m, both at m, each with lateral slots A/B at m and an outward offset to m; straight fences are rotated by angle and curved fences use radius m. The corresponding input uses one-hot encoding for fence type and presence at two locations, along with normalized angle values.
Figure 7.
Fence locations and encoding: Location 1 at m and Location 2 at m, both at m, each with lateral slots A/B at m and an outward offset to m; straight fences are rotated by angle and curved fences use radius m. The corresponding input uses one-hot encoding for fence type and presence at two locations, along with normalized angle values.
Figure 8.
Examples of fence configurations used in our dataset. Each panel shows a bird’s-eye view of a walkway (grey) with fences (white lines) on the left and/or right. The top row illustrates single-fence cases: (a) a straight fence on the right side, (b) a straight fence on the left side, (c) a curved fence on the right side, and (d) a curved fence on the left side. The bottom row shows the four possible two-fence combinations: (e) straight fences on both sides, (f) a straight fence on the right and a curved fence on the left, (g) a curved fence on the right and a straight fence on the left, and (h) curved fences on both sides. These eight scenarios cover the possible arrangements considered in the analysis.
Figure 8.
Examples of fence configurations used in our dataset. Each panel shows a bird’s-eye view of a walkway (grey) with fences (white lines) on the left and/or right. The top row illustrates single-fence cases: (a) a straight fence on the right side, (b) a straight fence on the left side, (c) a curved fence on the right side, and (d) a curved fence on the left side. The bottom row shows the four possible two-fence combinations: (e) straight fences on both sides, (f) a straight fence on the right and a curved fence on the left, (g) a curved fence on the right and a straight fence on the left, and (h) curved fences on both sides. These eight scenarios cover the possible arrangements considered in the analysis.
Figure 9.
Examples of final orientation distributions for different part–feeder configurations, represented as discrete PMFs with probability per bin, not CDFs. Axes: orientation angle in degrees on the x-axis and PMF, probability per bin, on the y-axis. The diversity in shapes, unimodal, multimodal, and uniform, highlights the complexity captured by the simulation data.
Figure 9.
Examples of final orientation distributions for different part–feeder configurations, represented as discrete PMFs with probability per bin, not CDFs. Axes: orientation angle in degrees on the x-axis and PMF, probability per bin, on the y-axis. The diversity in shapes, unimodal, multimodal, and uniform, highlights the complexity captured by the simulation data.
Figure 10.
Architecture of the standard fully connected regression model.
Figure 10.
Architecture of the standard fully connected regression model.
Figure 11.
VAE architecture. Part 1 is the part+feeder feature branch with 53 dimensions. Part 2 is the distribution branch with 360 CDF bins. The encoders share a latent space, and the decoder reconstructs both outputs.
Figure 11.
VAE architecture. Part 1 is the part+feeder feature branch with 53 dimensions. Part 2 is the distribution branch with 360 CDF bins. The encoders share a latent space, and the decoder reconstructs both outputs.
Figure 12.
Example predictions from Model 3 on the Test dataset (held-out configurations) for four representative cases, panels a–d. The model maintains good values on these held-out samples. Curves shown are circular CDFs, not PMFs, computed as unshifted cumulative sums from of the 360-bin PMF; PMFs are smoothed at dataset creation, wrapped Gaussian with , and no shared shift is applied. Model 3 outputs PMFs, and we plot their derived CDFs. Axes: orientation angle in degrees on the x-axis and CDF, unitless cumulative probability, on the y-axis.
Figure 12.
Example predictions from Model 3 on the Test dataset (held-out configurations) for four representative cases, panels a–d. The model maintains good values on these held-out samples. Curves shown are circular CDFs, not PMFs, computed as unshifted cumulative sums from of the 360-bin PMF; PMFs are smoothed at dataset creation, wrapped Gaussian with , and no shared shift is applied. Model 3 outputs PMFs, and we plot their derived CDFs. Axes: orientation angle in degrees on the x-axis and CDF, unitless cumulative probability, on the y-axis.
Figure 13.
Example predictions from Model 3 on the Parts dataset with unseen part geometries and only 3 unique parts for four representative cases. The significant performance drop highlights the difficulty of generalizing to new shapes. Curves shown are circular CDFs, not PMFs, computed as unshifted cumulative sums from of the 360-bin PMF; PMFs are smoothed at dataset creation, wrapped Gaussian with , and no shared shift is applied. Model 3 outputs PMFs, and we plot their derived CDFs. Axes: orientation angle in degrees on the x-axis and CDF, unitless cumulative probability, on the y-axis.
Figure 13.
Example predictions from Model 3 on the Parts dataset with unseen part geometries and only 3 unique parts for four representative cases. The significant performance drop highlights the difficulty of generalizing to new shapes. Curves shown are circular CDFs, not PMFs, computed as unshifted cumulative sums from of the 360-bin PMF; PMFs are smoothed at dataset creation, wrapped Gaussian with , and no shared shift is applied. Model 3 outputs PMFs, and we plot their derived CDFs. Axes: orientation angle in degrees on the x-axis and CDF, unitless cumulative probability, on the y-axis.
Figure 14.
Cross-convergence evaluation comparing within-level reconstruction (green, control) against partial-to-100% prediction (red). The within-level curve shows stable across all iteration levels, confirming the VAE architecture is sound. The cross-convergence curve shows that predicting fully converged CDFs from partial-iteration inputs yields poor results () at fewer than 30 iterations, with substantial improvement only above 60–70%. Error bars show ±1 SD across holdout configurations. This demonstrates that partial simulations lack sufficient information to predict final distributions until convergence is nearly complete.
Figure 14.
Cross-convergence evaluation comparing within-level reconstruction (green, control) against partial-to-100% prediction (red). The within-level curve shows stable across all iteration levels, confirming the VAE architecture is sound. The cross-convergence curve shows that predicting fully converged CDFs from partial-iteration inputs yields poor results () at fewer than 30 iterations, with substantial improvement only above 60–70%. Error bars show ±1 SD across holdout configurations. This demonstrates that partial simulations lack sufficient information to predict final distributions until convergence is nearly complete.
Table 1.
Key Simulation and Data Processing Parameters.
Table 1.
Key Simulation and Data Processing Parameters.
| Parameter |
Value |
Units/Notes |
| Simulator |
CoppeliaSim Edu 4.6 |
Platform [28] |
| Physics engine |
Bullet 2.78 |
Timestep 5 ms; 10 substeps [29] |
| Gravity |
Not explicitly set |
Scene default; not logged in API script |
| Contact solver iterations |
Not explicitly set |
Scene default; not logged |
| ERP/CFM |
Not explicitly set |
Engine defaults; not logged |
| Linear/angular damping |
Not explicitly set |
Scene defaults; not logged |
| Friction combine mode |
Not explicitly set |
Engine default; not logged |
| Restitution combine mode |
Not explicitly set |
Engine default; not logged |
| Conveyor speed |
0.10 |
m/s; chosen for stable transport and fence interaction in simulation; not calibrated to a specific feeder |
| Friction coefficients (Coulomb ) |
Parts 0.8; fences 0.5; belt 1.0 |
Static = kinetic; effective parameters for stable, friction-dominated motion; not calibrated [29] |
| Restitution coefficients |
|
Highly damped contacts to reduce bounce in Bullet [29] |
| Initial part orientation () |
(, , ) |
Flat on belt, random yaw |
| Simulation iterations per config. |
1000 |
Chosen based on JS divergence analysis (Figure 4) |
| Output angle range |
|
Degrees (simulator-native) |
| PMF smoothing |
Wrapped Gaussian |
Circular convolution;
|
| Part representation |
DFT |
15 largest coefficients (freq, mag, phase) |
| Feature vector (part) |
45 |
Standardized parameters (z-score) |
| Feeder representation |
8 |
2x [3 one-hot type + 1 scalar angle min–max scaled to on training split; angle set to 0 when type=None] |
| Fence geometry (straight) |
Fixed |
Rigid cuboid; dimensions fixed in CoppeliaSim scene (length/width/thickness not varied; values not logged) |
| Fence geometry (curved) |
|
Arc segment with fixed cross-section (length/width/thickness fixed in scene; values not logged); placement uses radius r
|
| Fence placement slots |
; ;
|
Locations 1/2 (downstream/upstream) and lateral A/B offsets |
| Conveyor implementation |
CoppeliaSim Conveyor model |
Built-in conveyor object with constant belt velocity; internal parameters not logged |
| Label representation (options) |
Full (360)PCA (170/85) |
Bin counts for PMF/CDF: PMF 360 or 170; CDF 360 or 85 |
Table 2.
Summary of Designed Regression Models
Table 2.
Summary of Designed Regression Models
| Model |
Label Type |
Label Repr. |
Architecture |
Loss Function Notes |
| 1 |
PMF |
PCA (170) |
Standard FC |
MSE |
| 2 |
PMF |
PCA (170) |
Branched FC |
MSE |
| 3 |
PMF |
Full (360) |
Standard FC |
MSE (CDF)+ KL (PMF)+ KS (CDF) |
| 4 |
PMF |
Full (360) |
Branched FC |
MSE (CDF)+ KL (PMF)+ KS (CDF) |
| 5 |
CDF |
PCA (85) |
Standard FC |
MSE (PMF)+ MSE (CDF) |
| 6 |
CDF |
PCA (85) |
Branched FC |
MSE (CDF) |
| 7 |
CDF |
Full (360) |
Standard FC |
MSE (PMF)+ KS (CDF) |
| 8 |
CDF |
Full (360) |
Branched FC |
MSE (PMF)+ KS (CDF) |
Table 3.
VAE Hyperparameters, CDF reconstruction
Table 3.
VAE Hyperparameters, CDF reconstruction
| Parameter |
Value |
| Latent Dimension |
35 |
| Learning Rate |
0.002 |
| Batch Size |
64 |
| Epochs |
120 |
| Optimizer |
Adam, weight decay 3e-5 |
| Gradient Clipping |
Max norm: 5.0 |
| Loss Weights: recon P+C, recon CDF, KL, KS penalty |
2, 8, 0.25, 20 |
| Gradient penalty weight,
|
4 |
| KL Annealing Schedule |
Weight grew linearly, capped at 0.25 |
Table 4.
Dataset Summary, total samples = 1,236.
Table 4.
Dataset Summary, total samples = 1,236.
| Dataset |
Samples |
Unique Parts |
Purpose |
Protocol |
| Main (Train) |
890 |
38 |
Model training |
– |
| Main (Test) |
158 |
38 |
Held-out configurations (random split) |
P1 |
| Fences |
78 |
18 |
Unseen fence configurations |
P1 |
| Parts |
110 |
3 |
Unseen part geometries |
P3 |
| Total (All) |
1,236 |
41 |
All datasets combined |
– |
Table 5.
Regression Model Performance on Training Data: and in deg, mean ± SD; configurations.
Table 5.
Regression Model Performance on Training Data: and in deg, mean ± SD; configurations.
| Model |
|
[deg] |
| Model 1 |
|
|
| Model 2 |
|
|
| Model 3 |
|
|
| Model 4 |
|
|
| Model 5 |
|
|
| Model 6 |
|
|
| Model 7 |
|
|
| Model 8 |
|
|
Table 6.
Regression Model Performance on Test Dataset: and in deg, mean ± SD; configurations.
Table 6.
Regression Model Performance on Test Dataset: and in deg, mean ± SD; configurations.
| Model |
|
[deg] |
| Model 1 |
|
|
| Model 2 |
|
|
| Model 3 |
|
|
| Model 4 |
|
|
| Model 5 |
|
|
| Model 6 |
|
|
| Model 7 |
|
|
| Model 8 |
|
|
Table 7.
Regression Model Performance on Fences Dataset, unseen feeder configurations: and in degrees, mean ± SD; configurations.
Table 7.
Regression Model Performance on Fences Dataset, unseen feeder configurations: and in degrees, mean ± SD; configurations.
| Model |
|
[deg] |
| Model 1 |
|
|
| Model 2 |
|
|
| Model 3 |
|
|
| Model 4 |
|
|
| Model 5 |
|
|
| Model 6 |
|
|
| Model 7 |
|
|
| Model 8 |
|
|
Table 8.
Regression Model Performance on Parts Dataset: and in deg, mean ± SD; configurations.
Table 8.
Regression Model Performance on Parts Dataset: and in deg, mean ± SD; configurations.
| Model |
|
[deg] |
| Model 1 |
|
|
| Model 2 |
|
|
| Model 3 |
|
|
| Model 4 |
|
|
| Model 5 |
|
|
| Model 6 |
|
|
| Model 7 |
|
|
| Model 8 |
|
|
Table 9.
VAE Reconstruction Performance, mean ± SD, unitless, on held-out Fences, , and Parts, , datasets. The VAE was trained only on the Main dataset, ; Fences and Parts were excluded from training. This table reports autoencoding reconstruction accuracy using both part+feeder features and CDF inputs, not conditional prediction from geometry alone.
Table 9.
VAE Reconstruction Performance, mean ± SD, unitless, on held-out Fences, , and Parts, , datasets. The VAE was trained only on the Main dataset, ; Fences and Parts were excluded from training. This table reports autoencoding reconstruction accuracy using both part+feeder features and CDF inputs, not conditional prediction from geometry alone.
| Dataset |
(Part+Feeder) |
(CDF) |
| Fences |
|
|
| Parts |
|
|
Table 10.
Delta-to-full correction performance (CDF ) at representative iteration levels. Cost saved is . Part split = unseen parts; config split = random configuration split with parts seen during training. Full 5% sweep results are provided in Supplementary Section S11.
Table 10.
Delta-to-full correction performance (CDF ) at representative iteration levels. Cost saved is . Part split = unseen parts; config split = random configuration split with parts seen during training. Full 5% sweep results are provided in Supplementary Section S11.
| Level |
Cost saved (%) |
Part split () |
Part baseline |
Config split () |
Config baseline |
| 5% |
95.0 |
0.82±0.22 |
-0.02±0.69 |
0.83±0.21 |
-0.02±0.69 |
| 10% |
90.0 |
0.82±0.21 |
-0.01±0.68 |
0.83±0.21 |
-0.01±0.69 |
| 25% |
75.0 |
0.84±0.18 |
0.03±0.62 |
0.84±0.20 |
0.03±0.62 |
| 50% |
50.0 |
0.86±0.16 |
0.31±0.35 |
0.86±0.18 |
0.31±0.36 |
| 75% |
25.0 |
0.96±0.07 |
0.89±0.06 |
0.95±0.06 |
0.89±0.06 |
| 90% |
10.0 |
0.99±0.01 |
0.99±0.01 |
0.99±0.01 |
0.99±0.01 |