| Class | Accuracy |
|---|---|
| Chair | 99.84% |
| Vase | 91.18% |
| Lamp | 97.44% |
| Visualization | True Class | Predicted Class | Analysis |
|---|---|---|---|
![]() |
Chair | Lamp | This chair has a vertical structure with thin legs and a tall back, which may resemble lamp-like features to the model. |
![]() |
Vase | Lamp | The narrow shape of this vase creates ambiguity with lamp structures, particularly given the vertical symmetry. |
![]() |
Lamp | Vase | This lamp has a wide base that shares geometric similarities with vase shapes, confusing the classifier. |
The PointNet classification model achieves great overall performance (98.32%), with particularly strong results on chairs (99.84%). The model struggles slightly more with vases (91.18%), likely because of their high shape variability and similarity to lamps.
Common failure modes include:
Analysis: This chair has complex, difficult to distinguish structures where the model struggles to maintain consistent part boundaries.
Analysis: Ambiguous part boundaries and uneven point density contribute to segmentation errors.
The segmentation model achieves strong overall performance (90.25%), with a clear distinction between easy and challenging cases. Best predictions (99%+ accuracy) occur on chairs with clear geometric boundaries between parts. The model successfully segments well-separated components.
Challenging predictions (44-46% accuracy) reveal common difficulties:
Procedure: Rotated point clouds around the z-axis by varying degrees (0, 45, 90) and evaluated segmentation accuracy.
| Rotation Angle | Accuracy | Change from 0 degree |
|---|---|---|
| 0 (Baseline) | 90.25% | - |
| 45 | 63.19% | -27.06% |
| 90 | 38.12% | -52.13% |
Potential improvements: Adding data augmentation with random rotations during training, or using rotation-invariant features (e.g., local reference frames, DGCNN's edge convolutions).
Procedure: Varied the number of points per object (500, 2,500, 10,000) to test how point cloud sparsity affects segmentation performance.
| Number of Points | Accuracy | Change from 10,000 |
|---|---|---|
| 10,000 (Baseline) | 90.25% | - |
| 2,500 | 90.22% | -0.03% (negligible) |
| 500 | 89.21% | -1.04% |
Key Findings:
Contrast with Experiment 1: Unlike rotation (58% accuracy drop at 90), point density reduction has minimal impact. This highlights that PointNet's learned features are more dependent on orientation than on dense sampling.
| Model | Overall Accuracy | Chair | Vase | Lamp |
|---|---|---|---|---|
| PointNet (Q1) | 98.32% | 99.84% | 91.18% | 97.44% |
| DGCNN (Q4) | 97.59% | 99.84% | 82.35% | 98.29% |
| Model | Overall Accuracy | Best Case | Worst Case |
|---|---|---|---|
| PointNet (Q2) | 90.25% | 99.62% | 44.40% |
| DGCNN (Q4) | 91.43% | 100.00% | 41.02% |
High-quality case: Both models perform well, but DGCNN achieves near-perfect segmentation. Note that I had to reduce the number of points for DGCNN due to memory constraints.
Challenging case: Both models struggle with complex geometry
Correct predictions: DGCNN successfully classifies objects across all categories
Unexpectedly, DGCNN performs slightly worse than PointNet (97.59% vs 98.32%).
DGCNN shows improvement: 1.18% accuracy gain (90.25% --> 91.43%) demonstrates the value of local geometric features for part-level tasks.