[null,null,["最后更新时间 (UTC):2025-08-07。"],[],[],null,["# Object Detection and Tracking\n=============================\n\nplat_ios plat_android \n\nWith ML Kit's on-device object detection and tracking API, you can localize\nand track in real time the most prominent objects in an image or live camera\nfeed. You can also optionally classify detected objects into one of several\ngeneral categories.\n\nObject detection and tracking with coarse classification is useful for building\nlive visual search experiences. Because object detection and tracking happens\nquickly and completely on the device, it works well as the front end of a longer\nvisual search pipeline. After you detect and filter objects, you can pass them\nto a cloud backend, such as [Cloud Vision Product Search](https://cloud.google.com/vision/product-search/docs/),\nor to a custom model, such as one you trained using\n[AutoML Vision Edge](/docs/ml-kit/automl-vision-edge).\n\n[iOS](/docs/ml-kit/ios/detect-objects)\n[Android](/docs/ml-kit/android/detect-objects)\n| This is a beta release of ML Kit for Firebase. This API might be changed in backward-incompatible ways and is not subject to any SLA or deprecation policy.\n\nKey capabilities\n----------------\n\n|------------------------------------|-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|\n| Fast object detection and tracking | Detect objects and get their location in the image. Track objects across images. |\n| Optimized on-device model | The object detection and tracking model is optimized for mobile devices and intended for use in real-time applications, even on lower-end devices. |\n| Prominent object detection | Automatically determine the most prominent object in an image. |\n| Coarse classification | Classify objects into broad categories, which you can use to filter out objects you're not interested in. The following categories are supported: home goods, fashion goods, food, plants, places, and unknown. |\n\nExample results\n---------------\n\n### Tracking the most prominent object across images\n\n|---|----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|\n| | |---------------------------|--------------------------------------------| | Tracking ID | 0 | | Bounds | (95, 45), (496, 45), (496, 240), (95, 240) | | Category | PLACE | | Classification confidence | 0.9296875 | |\n| | |---------------------------|--------------------------------------------| | Tracking ID | 0 | | Bounds | (84, 46), (478, 46), (478, 247), (84, 247) | | Category | PLACE | | Classification confidence | 0.8710938 | |\n| | |---------------------------|--------------------------------------------| | Tracking ID | 0 | | Bounds | (53, 45), (519, 45), (519, 240), (53, 240) | | Category | PLACE | | Classification confidence | 0.8828125 | |\n\nPhoto: Christian Ferrer \\[CC BY-SA 4.0\\]\n\n### Multiple objects in a static image\n\n| Object 0 ||\n|---------------------------|----------------------------------------------|\n| Bounds | (1, 97), (332, 97), (332, 332), (1, 332) |\n| Category | FASHION_GOOD |\n| Classification confidence | 0.95703125 |\n| Bounds | (186, 80), (337, 80), (337, 226), (186, 226) |\n| Category | FASHION_GOOD |\n| Classification confidence | 0.84375 |\n| Bounds | (296, 80), (472, 80), (472, 388), (296, 388) |\n| Category | FASHION_GOOD |\n| Classification confidence | 0.94921875 |\n| Bounds | (439, 83), (615, 83), (615, 306), (439, 306) |\n| Category | FASHION_GOOD |\n| Classification confidence | 0.9375 |"]]