发布 HN:OctaPulse(YC W26) – 鱼类养殖的机器人技术与计算机视觉
嗨,HN!我叫罗汉,与保罗一起是OctaPulse的联合创始人(<a href="https://www.tryoctapulse.com">https://www.tryoctapulse.com</a>)。我们正在为海鲜生产构建一个机器人层,首要任务是自动化鱼类检查。目前,我们已经在北美最大的鳟鱼生产商的首个生产基地投入使用。
你可能会想,我们怎么会在没有水产养殖或海洋产业背景的情况下进入这个领域。我们俩都来自沿海社区。我来自印度的果阿,保罗则来自马耳他和波多黎各。海鲜与我们的文化和社区息息相关。我们亲眼目睹了海洋所遭受的破坏,以及野生鱼类资源被捕捞到濒临灭绝的现状。我们还了解到,鱼类是全球近55%人口的主要蛋白质来源。尽管在美国的消费量并不大,但在全球范围内却是巨大的。然后我们发现,美国90%的海鲜依赖进口。什么?这听起来太荒谬了。这成为了我们创办这家公司的初衷。
保罗和我在卡内基梅隆大学的创业聚会上相遇。在像匹兹堡这样的内陆城市,很少有人关心海洋,更不用说想围绕海洋建立公司了。我们聚在一起讨论海洋技术和我们的兴趣,聊了整整三个小时。我对在海洋中进行开发产生了兴趣,因为这是最具挑战性的工程领域之一。保罗已经研究水产养殖几个月,发现同样的问题:这是一个3500亿美元的全球产业,但数据透明度却低得像一个仓库。经过那次谈话,我们知道我们想要一起在这个领域工作。
孵化场是生产的早期阶段,充满了劳动密集型的工作流程,非常适合自动化。农民需要测量鱼类的存量,以便做出喂养、繁殖和收获的决策,但鱼类在水下,处理时会感到压力。大多数农场仍然是手动取样。他们捕捞几打鱼,麻醉后放在桌子上逐一测量,然后推算出数十万的鱼群。这一过程每条鱼大约需要5分钟,数据也非常稀疏。
当我们看到这个过程时感到困惑。肯定有更好的方法。这是我们真正开始的起点。
不过,有一点需要注意。大多数机器人并不是为潮湿和湿润的环境设计的。盐水是任何机械设备的敌人,腐蚀问题非常棘手。更不用说水下计算机视觉需要穿透水的浑浊和颗粒。鱼类的运动不可预测,游动时会变形,遮挡现象时常发生。在不受控制的环境中进行校准也很棘手。用机器人处理活鱼是另一个尚未真正解决的挑战。鱼类滑腻、脆弱,容易受到压力。所有这些都要求所有材料必须符合食品安全标准。
在视觉方面,我们使用Luxonis OAK相机,它以紧凑的形式提供深度和RGB信息。机载的Myriad X VPU使我们能够直接在相机上运行轻量级推理,例如检测和跟踪,而无需不断通过USB发送原始帧。对于更重的工作负载,如分割和关键点提取,我们使用Nvidia Jetson。根据不同地点的功率和热量限制,我们在Orin Nano和Orin NX上进行了测试。
模型本身是基于CNN和变换器的架构。我们运行YOLO变体进行检测,定制分割头用于身体轮廓,以及关键点模型用于解剖标志。棘手的部分是让这些模型在边缘硬件上运行得足够快。我们根据部署目标使用TensorRT、OpenVINO和ONNX Runtime的组合。量化的过程也是一段旅程。TensorRT上的INT8量化为我们提供了所需的速度,但必须小心准确性下降,尤其是在分割输出中,边界精度至关重要。我们花了很多时间构建校准数据集,真实反映我们在农场看到的变化。光照在一天中变化,水的清晰度也会变化,鱼的密度各不相同。你的校准集需要捕捉到所有这些,否则你的量化模型在生产中就会崩溃。
在大多数农场没有WiFi的情况下,我们使用Starlink在偏远或海上位置进行连接。一切都在本地运行,连接可用时再进行同步。我们并没有将视频流传输到云端,所有推理都在设备上进行。
在幕后,我们一直在构建自己的内部工具,用于标注、任务分配和模型管理。早期我们尝试了现有的标注平台,但它们并不符合我们的工作流程。我们需要标注、训练管道和部署之间的紧密集成。因此,我们构建了自己的系统,可以将标注任务分配给注释员,跟踪进度,版本化数据集,并通过单个命令将模型推送到边缘设备。这并不华丽,但它让我们掌控一切,使迭代变得快速。当你试图在农场的数据收集、标注、训练、量化和部署之间闭环时,不能承受工具的碎片化。我们需要一个能够处理所有这些的系统。
在机器人方面,我们正在围绕现成组件构建定制外壳,并用软机器人抓手改造Delta机器人进行处理。在这种环境中,真空和典型的抓手驱动无法使用,因此我们使用可以安全处理鱼类而不损坏它们的柔性抓手。我们以Delta X S作为测试平台,并在验证湿润和潮湿环境中的运动学和负载要求后,评估是否转向工业Delta机器人或从头开始构建自己的机器人。末端执行器的设计仍在不断演变。鱼类根据物种和生长阶段有不同的大小和体型,因此我们需要能够适应的抓手。
目前,我们专注于水面上的操作。孵化场的表型、分类和质量检查。这些比完全水下部署更容易实现,启动成本也更低。我们的想法是,如果我们能够结合遗传数据、环境数据和表型图像,我们就能帮助农场识别哪些鱼应该繁殖,哪些应该淘汰。这就是选择性繁殖的起点。
让我们感到惊讶的是,早期只有极少数养殖鱼类经历过遗传改良程序。鸡的生长速度比1950年快了4倍,这得益于数十年的选择性繁殖。但大多数养殖鱼类基本上是野生基因。改善水产养殖遗传的机会巨大,但完全受限于测量。你无法改善你无法测量的东西,而农场迄今为止几乎无法大规模测量任何东西。
不过,行业是建立在信任之上的。我们处理的是活动物,农场对谁接近他们的鱼群非常谨慎。作为外部人士,这种信任必须通过努力赢得。保罗已经是可持续水产养殖联盟的未来领袖,但真正的转折点是在参加世界水产养殖学会(美国最大的会议)时。通过一个联系,他认识了我们第一个客户的首席遗传学家。这个关系转变为与北美最大的鳟鱼生产商的有偿试点。
我之前在ASML、Nvidia、特斯拉和丰田工作过。保罗在彭博社工作。我们在卡内基梅隆大学相遇,立刻知道我们想要解决这个问题,并将毕生精力投入其中。
我们非常希望听到任何在恶劣或不可预测环境中从事计算机视觉、在受限硬件上进行边缘部署,或用机器人温和适当地处理活动物的人的反馈。如果你在Jetson或OAK相机上运行推理,并对量化工作流程有看法,我们非常希望听到你们的成功经验。如果你有水产养殖的经验,我们也想知道有哪些问题是我们尚未遇到但应该考虑的。
丹告诉我们,你们都习惯于演示视频,但不幸的是,由于保密协议,我们无法分享。不过,这里有一张我们为表型和形态测量分析构建初始数据集的照片:[[<a href="https://drive.google.com/file/d/1z3oSlB8ed9hanrybzP24XTfjDJEobJua/view?usp=sharing" rel="nofollow">https://drive.google.com/file/d/1z3oSlB8ed9hanrybzP24XTfjDJE...</a>]]。
[[这是一个奇怪的行业,我们每周都在学习新的东西。如果你在边缘部署、湿润环境中的机器人或水产养殖方面有经验,我们非常希望听到你的看法。如果你对鱼类或我们的技术有任何问题,我们也很乐意在评论中深入讨论。期待听到这个社区的想法。]]
查看原文
Hi HN! My name is Rohan and, together with Paul, I’m the co-founder of OctaPulse (<a href="https://www.tryoctapulse.com/">https://www.tryoctapulse.com/</a>). We’re building a robotics layer for seafood production, starting with automated fish inspection. We are currently deployed at our first production site with the largest trout producer in North America.<p>You might be wondering how the heck we got into this with no background in aquaculture or the ocean industry. We are both from coastal communities. I am from Goa, India and Paul is from Malta and Puerto Rico. Seafood is deeply tied to both our cultures and communities. We saw firsthand the damage being done to our oceans and how wild fish stocks are being fished to near extinction. We also learned that fish is the main protein source for almost 55% of the world's population. Despite it not being huge consumption in America it is massive globally. And then we found out that America imports 90% of its seafood. What? That felt absurd. That was the initial motivation for starting this company.<p>Paul and I met at an entrepreneurship happy hour at CMU. In a landlocked city like Pittsburgh there aren't many people who care about the ocean, let alone want to build a company around it. We met to talk about ocean tech and our interests. It went on for three hours. I was drawn to building in the ocean because it is one of the hardest engineering domains out there. Paul had been researching aquaculture for months and kept finding the same thing: a $350B global industry with less data visibility than a warehouse. After that conversation we knew we wanted to work on this together.<p>Hatcheries, the early stage on-land part of production, are full of labor intensive workflows that are perfect candidates for automation. Farmers need to measure their stock for feeding, breeding, and harvest decisions but fish are underwater and get stressed when handled. Most farms still sample manually. They net a few dozen fish, anesthetize them, place them on a table to measure one by one, and extrapolate to populations of hundreds of thousands. It takes about 5 minutes per fish and the data is sparse.<p>When we saw this process we were baffled. There had to be a better way. This was the starting point that really kicked us off.<p>Here is the thing though. Most robots are not built to handle humid and wet environments. Salt water is the enemy of anything mechanical. Corrosion is such a pain to deal with. Don't get me started on underwater computer vision which has to parse through water turbidity and particles. Fish move unpredictably and deform while swimming. Occlusion is constant. Calibration is tricky in uncontrolled setups. Handling live fish with robotics is another challenge that hasn't really been solved before. Fish are slippery, fragile, and stress easily. All of this is coupled with the requirement that all materials must be food safe.<p>On the vision side we are using Luxonis OAK cameras which give us depth plus RGB in a compact form factor. The onboard Myriad X VPU lets us run lightweight inference directly on the camera for things like detection and tracking without needing to send raw frames over USB constantly. For heavier workloads like segmentation and keypoint extraction we bump up to Nvidia Jetsons. We have tested on the Orin Nano and Orin NX depending on power and thermal constraints at different sites.<p>The models themselves are CNN and transformer based architectures. We are running YOLO variants for detection, custom segmentation heads for body outlines, and keypoint models for anatomical landmarks. The tricky part is getting these to run fast enough on edge hardware. We are using a mix of TensorRT, OpenVINO, and ONNX Runtime depending on the deployment target. Quantization has been a whole journey. INT8 quantization on TensorRT gives us the speed we need but you have to be careful about accuracy degradation especially on the segmentation outputs where boundary precision matters. We spent a lot of time building calibration datasets that actually represent the variance we see on farms. Lighting changes throughout the day, water clarity shifts, fish density varies. Your calibration set needs to capture all of that or your quantized model falls apart in production.<p>There is no wifi at most of these farms so we are using Starlink for connectivity in remote or offshore locations. Everything runs locally first and syncs when connection is available. We are not streaming video to the cloud. All inference happens on device.<p>Behind the scenes we have been building our own internal tooling for labeling, task assignment, and model management. Early on we tried existing labeling platforms but they did not fit our workflow. We needed tight integration between labeling, training pipelines, and deployment. So we built our own system where we can assign labeling tasks to annotators, track progress, version datasets, and push models to edge devices with a single command. It is not fancy but it keeps everything under our control and makes iteration fast. When you are trying to close the loop between data collection on farm, labeling, training, quantization, and deployment you cannot afford to have fragmented tooling. We needed one system that handles all of it.<p>On the robotics side we are building custom enclosures around off the shelf components and modifying delta robots with soft robotics grippers for handling. Vacuum and typical gripper actuation will not work in this environment so we are using compliant grippers that can safely handle fish without damaging them. We started with the Delta X S as our test platform and are evaluating whether to move to industrial delta robots or build our own from scratch once we validate the kinematics and payload requirements in wet and humid environments. The end effector design is still evolving. Fish come in different sizes and body shapes depending on species and life stage so we need grippers that can adapt.<p>Right now we are focused on operations outside the water. Hatchery phenotyping, sorting, quality inspection. These are more accessible than full underwater deployment and cheaper to start with. The idea is that if we can combine genetics data, environmental data, and phenotypic imagery we can help farms identify which fish to breed and which to cull. This is where selective breeding starts.<p>Something that surprised us early on: only a tiny fraction of farmed fish species have been through genetic improvement programs. Chickens grow 4x faster than they did in 1950 because of decades of selective breeding. But most farmed fish are essentially wild genetics. The opportunity to improve aquaculture genetics is massive but it is completely bottlenecked on measurement. You cannot improve what you cannot measure, and farms can barely measure anything at scale so far.<p>The industry moves on trust though. We are dealing with live animals and farms are cautious about who they let near their stock. Coming from outside aquaculture, that trust had to be earned. Paul was already a Future Leader with the Coalition for Sustainable Aquaculture but the real turning point was attending World Aquaculture Society, the largest conference in the US. Through a connection of a connection he met the incoming lead geneticist at what became our first customer. That relationship turned into a paid pilot with the largest trout producer in North America.<p>I previously worked at ASML, Nvidia, Tesla, and Toyota. Paul worked at Bloomberg. We met at CMU and immediately knew that we wanted to tackle this problem and put our life's work into this.<p>We would love feedback from any of you who have worked on computer vision in harsh or unpredictable environments, edge deployment on constrained hardware, or gentle and appropriate handling of live animals with robotics. If you are running inference on Jetsons or OAK cameras and have opinions on quantization workflows we would love to hear what has worked for you. If you have aquaculture experience we are curious what problems we should be thinking about that we haven't encountered yet.<p>Dan told us you’re all used to demo videos but unfortunately we can’t share them due to NDAs. But here’s a photo of us building our initial dataset for phenotyping and morphometric analysis: [[<a href="https://drive.google.com/file/d/1z3oSlB8ed9hanrybzP24XTfjDJEobJua/view?usp=sharing" rel="nofollow">https://drive.google.com/file/d/1z3oSlB8ed9hanrybzP24XTfjDJE...</a>]].<p>[[This is a weird industry to be building in and we are learning something new every week. If you have experience with edge deployment, robotics in wet environments, or aquaculture itself we would love to hear your perspective. And if you just have questions about fish or the tech we are happy to go deep in the comments. Excited to hear what this community thinks.]]