The Good Tech Companies - The Hidden Bottlenecks of 3D Data Labeling

Episode Date: January 19, 2026

This story was originally published on HackerNoon at: https://hackernoon.com/the-hidden-bottlenecks-of-3d-data-labeling. A “simple” process of the 3D labeling often ...includes challenges and hidden bottlenecks. How can they be solved, and what is the optimal solution for them? Check more stories related to machine-learning at: https://hackernoon.com/c/machine-learning. You can also check exclusive content about #ai, #lidar, #3d, #3d-labeling, #3d-point-cloud-navigation, #point-cloud, #data-annotation-services, #good-company, and more. This story was written by: @keymakr. Learn more about this writer by checking @keymakr's about page, and for more stories, please visit hackernoon.com. 3D labeling unlocks powerful capabilities for autonomous systems across various industries, including automotive, robotics, construction, and healthcare. Point-cloud data is inherently unstable: reflections from glass or wet surfaces, weather-induced noise, and constantly moving objects can distort the scene.

Transcript
Discussion (0)
Starting point is 00:00:00 This audio is presented by Hacker Noon, where anyone can learn anything about any technology. The hidden bottlenecks of 3D data labeling by Keymaker. For engineers who perfected 2D data pipelines, moving to 3D can be a shock. What was once a solved problem of drawing 2D boxes has become a complex battle against sparse point clouds, clunky visualization, and ambiguous classifications. A simple process often includes numerous challenges and hidden bottlenecks. How can they they be solved and what is the optimal solution for them? Let's try to find it in the field by exploring the expertise and projects of the leading labeling provider and 3D platform. Where the bottlenecks arise 3D labeling unlocks powerful capabilities for autonomous systems
Starting point is 00:00:45 across various industries, including automotive, robotics, construction, and healthcare. So, where do a unique set of challenges arise? Point cloud data is inherently unstable, Reflections from glass or wet surfaces, weather-induced noise, and constantly moving objects can distort the scene or create phantom structures that never existed. Large-scale scans add additional complexity, aligning sensors, synchronizing frames, and maintaining geometric consistency across massive data sets requires meticulous control. Even more importantly, three-d annotation is more than marking points. It's about correctly interpreting context.
Starting point is 00:01:24 A litter system doesn't know that a storefront is. transparent, that a CYC list may suddenly change posture, or that two nearly identical objects should be understood differently in the real world. Another frequent bottleneck appears long before annotation even begins, unclear goals. The level of detail, the attributes to capture, and the semantic relationships between objects must all be defined up front. Whether a bicycle needs detailed segmentation or a single cuboid with a movement vector shapes the entire architecture of the model. individually, these challenges are typical for any complex point cloud project. Together, they can slow down or completely block an AI pipeline.
Starting point is 00:02:04 All this can be solved by a solution built on three pillars, experienced specialists, proper tooling, and clearly defined requirements. Joint Laboratory of Humans and Powerful Tools The announced partnership between Keymaker, a team specializing in high-precision labeling and segments. I, a platform designed for enterprise grade labeling of 3D, demonstrates how the combination of advanced tooling and human expertise elevates 3D data workflows and solves hidden bottlenecks. Working across 3D projects of different complexity and scale, the keymaker and segments, ITEMs share their most interesting cases and creative solutions to
Starting point is 00:02:41 challenging problems. Solving weather-related point cloud distortions one of the practical examples of solving bottlenecks came from weather-related noise. During a litter-based infrastructure mapping project, heavy fog and drizzle caused unstable point density, making it difficult to distinguish real objects from atmospheric artifacts. Segments, eyes built-in filters handled the first layer of cleanup by removing low intensity and isolated points. However, the weather distortions were too heterogeneous for automatic tools alone. Fog created diffuse point clusters, drizzle generated sharp but random, sparkles, and reflective road surfaces added another layer of complexity. To address this, the Keymaker team applied a multi-step review workflow.
Starting point is 00:03:26 Pattern-based visual analysis, annotators manually identified recurring noise signatures, fog patches, reflective glare, moisture artifacts, iterative threshold tuning. Segments, I engineers adapted filtering rules to environmental interference rather than generic noise. Layered cleanup. The dataset was filtered, reviewed, corrected, and then passed through an adjusted filter a second time. Cross-checking in 3D space. Operators validated and cleaned regions in multiple angles and depths to ensure legitimate objects were preserved. This human machine loop dramatically improved the dataset. Even more importantly, it produced a repeatable workflow for weather-affected scans. A process both teams now use as a template when working
Starting point is 00:04:10 with litter captured in in ideal conditions. Greater than, Keymaker's team makes full use of our platform to optimize speed and quality greater than across complex projects. Their proactive communication and expert feedback greater than help us continuously improve the platform. It's always great to see the greater than results they're achieving for some of the customers we support together. Greater than greater than Otto DeBals, CEO at segments. AI restoring time synchronization in custom litter formats. Another case involved a technical bottleneck related to time-based point cloud data. Unlike video, where frames are discrete and easy to align, point CloudSauri measured over time, meaning not a specific frame
Starting point is 00:04:51 but a continuous recording in seconds. The client had a custom data format, and during upload, the frame to timestamp alignment was overlooked, causing the file to display incorrectly. As a result, the platform couldn't recognize or process the dataset. The challenge was resolved jointly by the keymaker technical team and segments. I-engineers, who adjusted the time-mapping logic and restored full compatibility. This collaboration salvaged the dataset and led to improvements in the platform's time synchronization capabilities for future projects. Greater than, we're always looking for creative ways to use our partners' tools, as our greater than clients often request non-standard and complex tasks, says Zoya Boyko, PM at Greater Than Keymaker. Working with unconventional formats and challenging cases
Starting point is 00:05:38 helps greater than segments improve the platform, creating a better environment for everyone. So, greater than we grow together, and our team finds its tools indispensable. Thus, Keymaker contributes the human component, attentiveness, creativity, and a deep understanding of how real-world data behaves. Segment brings the streamlining element, powering workflows, engineering solutions, and making labeling possible. When the annotation team tests real, field, cases on the platform, both companies create an environment where the tool adapts to the user end, in turn, shapes the workflows accordingly. Resolving structural breaks in 3D road markings one more bottleneck emerged in projects requiring
Starting point is 00:06:20 highly accurate 3D road markings. In 2D workflows, a line can be interrupted and continued without affecting the overall structure, but in 3D, this is dangerous. Even a tiny artificial gap can be misinterpreted by an autonomous system as a lane ending. To address this, the keymaker team developed a workflow based on continuous polylines with automated overlap removal. Each road marking was drawn as a single continuous object, after which segments. I automatically removed only the segments that fell within overlap zones. This prevented the model from reading overlaps as structural brakes, significantly improving consistency. Greater than, in 3D, you can't afford accidental gaps. A single break can distort the greater than geometry of the entire scene and mislead an autonomous system.
Starting point is 00:07:08 Our goal was to greater than create a workflow where continuity is guaranteed by design, not left to greater than chance, Zoya Boyko, PM at Keymaker. The process was further strengthened by working inside an aggregated 3D scene using segments. Eyes merge mode. Instead of thousands of disconnected frames, annotators worked with a unified, dense representation of the entire route, complete with edges, splits, merges, curbs, signs, and guardrails. This holistic view made road geometry clearer, transitions more logical, and lines smoothing far more accurate. After building the global skeleton of the annotation in merge mode, operators switched back to frame-level editing only to refine overlaps without altering the underlying geometry. This kept datasets compact and avoided errors common inextensive multi-frame edits. What began as a structural bottleneck became a scalable method now reused across many 3D litter projects. All these cases from
Starting point is 00:08:07 keymaker and segments, I can be seen as examples of a joint laboratory where technology evolves in tandem with human expertise. While 3D labeling is often viewed as a purely technical process, in reality, it reflects something much deeper, teaching machines to understand the real world. Every point cloud captures a fragment of reality, and it is the human operator who determines what is meaningful, what is noise, and what truly exists in the scene. Together, these cases show that the future of 3D data work lies in Ahuman in the loop model, where people guide the system, shape the tools through real field challenges, and ultimately ensure reliability, accuracy, and trust in the final output. Thank you for listening to this Hackernoon story, read by artificial
Starting point is 00:08:51 intelligence. Visit Hackernoon.com to read, write, learn and publish.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.