ONVIF VideoAnalytics Service Spec v210

  • Upload
    12o0n3y

  • View
    232

  • Download
    0

Embed Size (px)

Citation preview

  • 7/30/2019 ONVIF VideoAnalytics Service Spec v210

    1/27

  • 7/30/2019 ONVIF VideoAnalytics Service Spec v210

    2/27

  • 7/30/2019 ONVIF VideoAnalytics Service Spec v210

    3/27

  • 7/30/2019 ONVIF VideoAnalytics Service Spec v210

    4/27

  • 7/30/2019 ONVIF VideoAnalytics Service Spec v210

    5/27

  • 7/30/2019 ONVIF VideoAnalytics Service Spec v210

    6/27

  • 7/30/2019 ONVIF VideoAnalytics Service Spec v210

    7/27

    ONVIF 7 Video Analytics. Ver. 2.10

    The temporal and spatial relation of scene elements with respect to the selected video sourceis discussed in sections 5.1.2.1 and 5.1.2.2. The appearance and behaviour of trackedobjects is discussed in section 5.1.3.1. Interactions between objects like splits and mergesare described in section 5.1.3.2.

    A PTZ device can put information about the Pan, Tilt and Zoom at the beginning of a frame,allowing a client to estimate the 3D coordinates of scene elements. Next, the imagecoordinate system can be adapted with an optional transformation node which is described inthe next subsection. Finally, multiple object descriptions can be placed and their associationcan be specified within an ObjectTree node.Below, the definitions are included forconvenience1:

    . . .

    . . .

    Subsection 5.1.2.1 describes how frames processed by the video analytics algorithm arereferenced within the video analytics stream.

    5.1.2.1 Temporal Relatio n

    Since multiple scene elements can be extracted from the same image, scene elements arelisted below a frame node which establishes the link to a specific image from the video input.The frame node contains a mandatory UtcTime attribute. This UtcTime timestamp shall enablea client to map the frame node exactly to one video frame. For example,. the RTP timestampof the corresponding encoded video frame shall result in the same UTC timestamp afterconversion. The synchronization between video and metadata streams is further described inthe ONVIF Streaming Specification.

    Example:

    . . .. . .

    . . .

    5.1.2.2 Spatial Relation

    Most scene elements refer to some part in an image from which information has beenextracted. For instance, when tracking objects over time, their position within each frame shallbe specified. These positions shall relate to a Coordinate System. The default coordinate

    system is shown in Figure 2.

    1 Please note that the schema is included here for information only. [ONVIF Schema] containsthe normative schema definition.

  • 7/30/2019 ONVIF VideoAnalytics Service Spec v210

    8/27

  • 7/30/2019 ONVIF VideoAnalytics Service Spec v210

    9/27

    ONVIF 9 Video Analytics. Ver. 2.10

    yyy

    xxx

    y

    x

    tsv

    tsv

    t

    t

    '

    '

    , , where and replace the top-most Coordinate

    System.

    yy

    xx

    y

    x

    su

    su

    s

    s

    '

    '

    '

    '

    y

    x

    t

    t

    '

    '

    y

    x

    s

    s

    For example, the coordinates of the scene description are given in a frame coordinate system,where the lower-left corner has coordinates (0,0) and the upper-right corner coordinates(320,240). The Frame Node resembles the following code where the scaling is set to thedoubled reciprocal of the frame width and the frame height:

    . . .

    5.1.3 Scene Elements

    This section focuses on scene elements generated by object tracking algorithms and definesobject handling and object shapes for them.

    Frames where no objects have been detected can be skipped within the Scene Description tosave bandwidth, as long as the last frame in the Scene Description is empty as well. It isrecommended that the device regularly sends the Scene Description even if it is empty, inorder to indicate that the analytics engine is operational. The device shall send a SceneDescription if a SynchronizationPoint is requested for the corresponding stream.

    When the receiver of a Scene Description receives an empty frame, the receiver should

    assume that all subsequent frames are empty as well until the next non-empty frame isreceived. When the last received frame is non-empty, the receiver should assume that adescription of the next processed frame will be transmitted.

    5.1.3.1 Objects

    Objects are identified via their ObjectID. Features relating to one particular object arecollected in an object node with the corresponding ObjectID as an attribute. Associations ofobjects, like Object Renaming, Object Splits, Object Merges and Object Deletions areexpressed in a separate ObjectTree node. An ObjectID is implicitly created with the firstappearance of the ObjectID within an object node3.

    . . .

    3 Please note that the schema is included here for information only. [ONVIF Schema] contains the normativeschema definition.

  • 7/30/2019 ONVIF VideoAnalytics Service Spec v210

    10/27

  • 7/30/2019 ONVIF VideoAnalytics Service Spec v210

    11/27

    ONVIF 11 Video Analytics. Ver. 2.10

    . . .

    . . .

    5.1.3.2 Object Tree

    When two objects come too close to each other, such that the video analytics can no longertrack them individually, an object Merge should be signalled by adding a merge node to theObjectTree node of the frame node. The merge node contains a From node listing themerging ObjectIds and a To node containing the ObjectId. The merged object is used in futureframes as the tracking ID. If the video analytics algorithm detects that one object is occludingthe others and is able to track this object further, the occluding object should be put in the Tonode.

    The separation of objects is indicated by a Split node. In this case, the From node contains asingle ObjectId representing the object which is split in the current frame. The objectsseparating from this split object are listed in the To node. The ObjectId of the From node canreappear in the To node, if this object did occlude the others and the video analytics algorithmwas able to track this object during the occlusion.

    An object does not need to be involved in a merge operation in order to be part of a splitoperation. For example, if an object is moving together with a person, and the person leavesthe object somewhere, the object might be detected the first time by the video analytics whenthe person moves away from the object left behind. In such cases, the first appearance of theobject can be combined with a Split operation.

    When a merged object reappears as an object node in a later frame without a split indication,then this object is implicitly split. The video analytics algorithm, however, could not determinewhere the split object came from.

  • 7/30/2019 ONVIF VideoAnalytics Service Spec v210

    12/27

    ONVIF 12 Video Analytics. Ver. 2.10

    A video analytics algorithm can track and remember a limited number of objects. In order toindicate that a certain object has been removed from the memory of the algorithm andtherefore never appear again, the Scene Description can contain a Delete node within theObjectTree node.

    If the video analytics algorithm can not decide during a Split operation the identity of an object,it should use a new ObjectId. When the algorithm has collected sufficient evidence for theidentity of this object, it can change the ObjectId via the Rename operation. The Renameoperation can also be used when an object reenters the scene and the true identity isdiscovered after some time.

    A deleted ObjectId shall not be reused within the Scene Description until the ObjectIdcontainer has wrapped around.

    Example:

    . . .

    . . .

    . . .

    . . .

    . . .

    . . .

    5.1.3.3 Shape desc riptor

    Shape information shall be placed below the optional shape node of in an object appearancenode. If present, the shape node holds information where the object under consideration has

  • 7/30/2019 ONVIF VideoAnalytics Service Spec v210

    13/27

    ONVIF 13 Video Analytics. Ver. 2.10

    been detected in the specified frame. A shape node shall at least contain two nodesrepresenting the BoundingBox and the CenterOfGravity of the detected object.

    The coarse BoundingBox is further refined with additional child nodes, each representing a

    shape primitive. If multiple shape primitives are present, their union defines the objects shape.In this specification, a generic polygon descriptor is provided.

    Polygons that describe the shape of an object shall be simple polygons defined by a list ofpoints.

    Two consecutive points (where the last point is connected with the first one) in the list definea line segment. The order of the points shall be chosen such that the enclosed object regioncan be found on the left-hand side all line segments. The polyline defined by the list of pointsshall not be self-intersecting.

    Example:

  • 7/30/2019 ONVIF VideoAnalytics Service Spec v210

    14/27

    ONVIF 14 Video Analytics. Ver. 2.10

    5.2 Rule interf ace

    A XML structure is introduced in Section 5.2.1 to communicate the configuration of rules.Section 5.2.2 specifies a language to describe the configuration of a specific rule type. Insection 5.2.3 two rules are specified that should be supported by a device implementing aRule Engine. Section 5.2.4 introduces operations to manage rules. If the device supports aRule Engine, it shall implement the complete rule Interface.

    5.2.1 Rule representation

    The configuration of a rule has two required attributes: one specifies the name and the otherspecifies the type of the rule. The different configuration parameters are listed below theparameters element of the rule element. Each parameter is either a SimpleItem or anElementItem. The name attribute of each item shall be unique within the parameter list.SimpleItems have an additional Value attribute containing the value of the parameter. Thevalue of ElementItems is given by the child element of the ElementItem. It is recommended torepresent as many parameters as possible by SimpleItems.

    The following example shows a complete video analytics configuration containing two rules:

    . . .

    5.2.2 Rule descri ptio n language

    The description of a rule contains the type information of all parameters belonging to a certainrule type and the description of the output produced by such a rule. The output of the RuleEngine is Events which can either be used in an Event Engine or be subscribed to by a client.

    The parameters of a certain rule type are listed below the ParameterDescription element. All

    parameters are either SimpleItems or ElementItems and can be described by either aSimpleItemDescription or an ElementItemDescription. Both ItemDescriptions contain a nameattribute to identify the parameter and a Type attribute to reference a specific XML schematype. In case of the SimpleItemDescription, the type attribute shall reference a SimpleType

  • 7/30/2019 ONVIF VideoAnalytics Service Spec v210

    15/27

    ONVIF 15 Video Analytics. Ver. 2.10

    schema definition. In case of the ElementItemDescription, the Type attribute shall reference aglobal element declaration of an XML schema.

    The output produced by this rule type is described in multiple MessageDescription elements.

    Each MessageDescription contains a description of the message payload according to theMessage Description Language detailed in the ONVIF Core specification. Additionally, theMessageDescription shall contain a ParentTopic element naming the Topic a client has tosubscribe to in order to receive this specific output. The topic shall be specified as a ConcreteTopic Expression.

    Section 5.2.3 demonstrates the usage of the Rule Description Language. Below, thedefinitions are included for convenience4:

    . . .

    5.2.3 Specif ied Rules

    The following rules apply to static cameras. In case of a PTZ device, image-based rulesshould contain an additional ElementItem. The ElementItem identifies the position of thedevice for which the rule has been setup. The corresponding ElementItemDescriptionresembles the following:

    4 Please note that the schema is included here for information only. [ONVIF Schema] containsthe normative schema definition.

  • 7/30/2019 ONVIF VideoAnalytics Service Spec v210

    16/27

    ONVIF 16 Video Analytics. Ver. 2.10

    5.2.3.1 LineDetector

    The LineDetector is defined by a non-intersecting simple polyline. If an Object crosses thepolyline in the specified direction, the Rule Engine sends a Crossed event containing thename of the LineDetector and a reference to the object which has crossed the line. Asdirections, one can select between Left, Right, and Any, where directions Left and Right referto the direction walking along the line from the first point to the second point and are theprohibited directions.

    The LineDetector resembles the following code using the Rule Description Language, detailedin the previous section:

    t ns1: Rul eEngi ne/ Li neDet ect or/ Cr ossed

    The code above defines two parameters, Segments and Direction, and produces one Eventattached to the topic tns1:RuleEngine/LineDetector/Crossed.

    5.2.3.2 FieldDetector

    A FieldDetector is defined by a simple non-intersecting polygon. The FieldDetectordetermines if each object in the scene inside or outside the polygon. This information is putinto a property.

    The FieldDetector resembles the following code, using the Rule Description Languagedetailed in the previous section:

  • 7/30/2019 ONVIF VideoAnalytics Service Spec v210

    17/27

    ONVIF 17 Video Analytics. Ver. 2.10

    t ns1: Rul eEngi ne/ Fi el dDet ect or / Obj ect sI nsi de

    From the Inside property, a client can derive the Entering and the Leaving parameters of thedetector. A client can simulate Entering and Leaving events by adding a MessageContentFilter to the subscription, which lets only ObjectsInside messages pass, where the IsInsideItem is set to true resp. false.

    5.2.4 Operations on rules

    If the device supports a Rule Engine as defined by ONVIF, then it shall implement thefollowing operations to manage rules. The Create/Delete/Modify operations are atomic,meaning that either all modifications can be processed or the complete operation shall fail.

    5.2.4.1 Get Suppo rted rul es

    The device shall indicate the rules it supports by implementing the subsequent operation. Itreturns a list of rule descriptions according to the Rule Description Language described inSection 5.2.2. Additionally, it contains a list of URLs that provide the location of the schemafiles. These schema files describe the types and elements used in the rule descriptions. If ruledescriptions reference types or elements of the ONVIF schema file, the ONVIF schema fileshall be explicitly listed.

    Table 1: GetSupportedRules c ommand

    GetSupportedRules Access Class: READ_MEDIA

    Message name Descript ion

    GetSupportedRulesRequest The request message contains the VideoAnalyticsConfigurationTokenfor which the supported rules should be listed.

    tt:ReferenceToken ConfigurationToken [1][1]

    GetSupportedRulesResponse The response contains the supported rules.

    tt:SupportedRules SupportedRules [1][1]

    Fault codes Description

    env:Sender

    ter:InvalidArgValter:NoConfig

    VideoAnalyticsConfiguration does not exist.

    5.2.4.2 Get Rul es

    The following operation retrieves the currently installed rules:

    Table 2: GetRules com mand

    GetRules Access Class: READ_MEDIA

    Message name Descript ion

    GetRulesRequest The request message specifies the VideoAnalyticsConfigurationTokenfor which the rules should be reported.

    tt:ReferenceToken ConfigurationToken [1][1]

  • 7/30/2019 ONVIF VideoAnalytics Service Spec v210

    18/27

    ONVIF 18 Video Analytics. Ver. 2.10

    GetRulesResponse The response is a list of installed rules for the specified configuration.

    tt:Config Rule [0][unbounded]

    Fault codes Descriptionenv:Senderter:InvalidArgValter:NoConfig

    The VideoAnalyticsConfiguration does not exist.

    5.2.4.3 Create rul es

    The following operation adds rules to a VideoAnalyticsConfiguration. If all rules can not becreated as requested, the device responds with a fault message.

    Table 3: CreateRules co mmand

    CreateRules Access Class: ACTUATEMessage name Descript ion

    CreateRulesRequest The request message specifies the VideoAnalyticsConfigurationTokento which the listed Rules should be added.

    tt:ReferenceToken ConfigurationToken [1][1]tt:Config Rule [1][unbounded]

    CreateRulesResponse This is an empty message.

    Fault codes Description

    env:Senderter:InvalidArgValter:NoConfig

    The VideoAnalyticsConfiguration does not exist.

    env:Senderter:InvalidArgVal

    ter:InvalidRule

    The suggested rules configuration is not valid on the device.

    env:Senderter:InvalidArgVal

    ter:RuleAlreadyExistent

    The same rule name exists already in the configuration.

    enc:Receiverter:Action

    ter:TooManyRules

    There is not enough space in the device to add the rules to theconfiguration.

    env:Receiverter:Action

    ter:ConfigurationConflict

    The device cannot create the rules without creating a conflictingconfiguration.

    5.2.4.4 Modi fy Rules

    The following operation modifies multiple rules. If all rules can not be modified as requested,the device responds with a fault message.

    Table 4: ModifyRules command

    ModifyRules Access Class: ACTUATEMessage name Descript ion

    ModifyRulesRequest The request message specifies the VideoAnalyticsConfigurationTokenfor which the listed Rules should be modified.

  • 7/30/2019 ONVIF VideoAnalytics Service Spec v210

    19/27

    ONVIF 19 Video Analytics. Ver. 2.10

    tt:ReferenceToken ConfigurationToken [1][1]tt:Config Rule[1][unbounded]

    ModifyRulesResponse This is an empty message.

    Fault codes Description

    env:Senderter:InvalidArgValter:NoConfig

    The VideoAnalyticsConfiguration does not exist.

    env:Senderter:InvalidArgVal

    ter:InvalidRule

    The suggested rules configuration is not valid on the device.

    env:Senderter:InvalidArgs

    ter:RuleNotExistent

    The rule name or names do not exist.

    enc:Receiverter:Actionter:TooManyRules

    There is not enough space in the device to add the rules to theconfiguration.

    env:Receiverter:Action

    ter:ConflictingConfig

    The device cannot modify the rules without creating a conflictingconfiguration.

    5.2.4.5 Delete Rules

    The following operation deletes multiple rules. If all rules can not be deleted as requested, thedevice responds with a fault message.

    Table 5: DeleteRules com mand

    DeleteRules Access Class: ACTUATE

    Message name Descript ion

    DeleteRulesRequest The request message specifies the VideoAnalyticsConfigurationTokenfrom which the listed Rules should be removed.

    tt:ReferenceToken ConfigurationToken [1][1]xs:string RuleName [1][unbounded]

    DeleteRulesResponse The response is an empty message.

    Fault codes Description

    env:Senderter:InvalidArgValter:NoConfig

    The VideoAnalyticsConfiguration does not exist.

    env:Receiverter:Action

    ter:ConflictingConfig

    The device cannot delete the rules without creating a conflictingconfiguration.

    env:Senderter:InvalidArgs

    ter:RuleNotExistent

    The rule name or names do not exist.

    5.3 Analyt ics Modules Interface

    Section 5.3.1 defines an XML structure that communicates the configuration of AnalyticsModules. Section 5.3.2 defines the language that describes the configuration of a specificanalytics module. Section 5.3.3 defines the operations required by the analytics modules

  • 7/30/2019 ONVIF VideoAnalytics Service Spec v210

    20/27

    ONVIF 20 Video Analytics. Ver. 2.10

    Interface. If the device supports an analytics engine as defined by ONVIF, it shall implementthe complete Analytics Modules Interface.

    5.3.1 Analytics module configur ation

    The analytics module configuration is identical to the rule configuration, described in section5.2.1. The following example shows a possible configuration of a vendor-specificObjectTracker. This tracker allows configuration of the minimum and maximum object sizewith respect to the processed frame geometry.

    . . .

    5.3.2 Analyt ics Modul e Descr ipt ion Language

    The Analytics Module reuses the Rule Description Language, described in Section 5.2.2. Thefollowing AnalyticsModuleDescription element replaces the RuleDescription element:

    Similar to rules, analytics modules produce events and shall be listed within the analyticsmodule description. The subsequent description corresponds to the example of the previoussection. The example module produces a SceneTooCrowded Event when the scene becomestoo complex for the module.

    t ns1: Vi deoAnal yt i cs/ nn: Obj ect Tracker/ SceneTooCr owded

    5.3.3 Operatio ns on Analyt ics Modules

    If the device supports an analytics engine as defined by ONVIF, it shall support thesubsequent operations to manage analytics modules. The Create/Delete/Modify operationsshall be atomic, all modifications can be processed or the complete operation shall fail.

  • 7/30/2019 ONVIF VideoAnalytics Service Spec v210

    21/27

    ONVIF 21 Video Analytics. Ver. 2.10

    5.3.3.1 GetSupportedAnalyticsModules

    The device indicates the analytics modules it supports by implementing theGetSupportedAnalysticsModule operation. It returns a list of analytics modules according tothe Analytics Module Description Language, described in section 5.2.2. Additionally, itcontains a list of URLs that provide the location of the schema files. These schema filesdescribe the types and elements used in the analytics module descriptions. If the analyticsmodule descriptions reference types or elements of the ONVIF schema file, theONVIFschema file shall be explicitly listed.

    Table 6: GetSupportedAnalyticsModules command

    GetSupportedAnalyticsModules Access Class: READ_MEDIA

    Message name Descript ion

    GetSupportedAnalyticsModulesRequest

    The request message contains theVideoAnalyticsConfigurationToken for which the supported

    analytics modules should be listed.

    tt:ReferenceToken ConfigurationToken [1][1]

    GetSupportedAnalyticsModulesResponse

    The response contains the supported analytics modules.

    SupportedAnalyticsModules [1][1]

    Fault codes Description

    env:Senderter:InvalidArgs

    ter:NoConfig

    VideoAnalyticsConfiguration does not exist.

    5.3.3.2 GetAn alytic s Modules

    The following operation retrieves the currently installed analytics modules:

    Table 7: GetAnalyticsModules command

    GetAnalyticsModules Access Class: READ_MEDIA

    Message name Descript ion

    GetAnalyticsModulesRequest The request message specifies the

    VideoAnalyticsConfigurationToken for which the analyticsmodules should be reported.

    tt:ReferenceToken ConfigurationToken [1][1]

    GetAnalyticsModulesResponse The response is a list of installed analytics modules for thespecified configuration.

    tt:ConfigAnalyticsModule [0][unbounded]

    Fault codes Description

    env:Senderter:InvalidArgster:NoConfig

    The VideoAnalyticsConfiguration does not exist.

  • 7/30/2019 ONVIF VideoAnalytics Service Spec v210

    22/27

    ONVIF 22 Video Analytics. Ver. 2.10

    5.3.3.3 CreateAnalyti cs Modul es

    The following operation adds analytics modules to a VideoAnalyticsConfiguration. If allanalytics modules can not be created as requested, the device responds with a fault message.

    Table 8: CreateAnalyticsModules command.

    CreateAnalyticsModules Access Class: ACTUATE

    Message name Descript ion

    CreateAnalyticsModulesRequest The request message specifies theVideoAnalyticsConfigurationToken to which the listed AnalyticsModules should be added.

    tt:ReferenceToken ConfigurationToken [1][1]tt:ConfigAnalyticsModule [1][unbounded]

    CreateAnalyticsModulesResponse This is an empty message.

    Fault codes Description

    env:Senderter:InvalidArgster:NoConfig

    The VideoAnalyticsConfiguration does not exist.

    env:Senderter:InvalidArgster:NameAlreadyExistent

    The same analytics module name exists already in theconfiguration.

    enc:Receiverter:Action

    ter:TooManyModules

    There is not enough space in the device to add the analyticsmodules to the configuration.

    env:Receiverter:Action

    ter:ConfigurationConflict

    The device cannot create the analytics modules without creatinga conflicting configuration.

    env:Senderter:InvalidArgVal

    ter:InvalidModule

    The suggested module configuration is not valid on the device.

    5.3.3.4 Modify Analyt ics Modules

    The following operation modifies multiple analytics modules. If all analytics modules can notbe modified as requested, the device respond with a fault message.

    Table 9: ModifyA nalyticsModules c ommand

    ModifyAnalyticsModules Access Class: ACTUATE

    Message name Descript ion

    ModifyAnalyticsModulesRequest The request message specifies theVideoAnalyticsConfigurationToken for which the listed analyticsmodules should be modified.

    tt:ReferenceToken ConfigurationToken [1][1]tt:ConfigAnalyticsModule [1][unbounded]

    ModifyAnalyticsModulesResponse The response is an empty message.

    Fault codes Description

  • 7/30/2019 ONVIF VideoAnalytics Service Spec v210

    23/27

    ONVIF 23 Video Analytics. Ver. 2.10

    env:Senderter:InvalidArgster:NoConfig

    The VideoAnalyticsConfiguration does not exist.

    env:Sender

    ter:InvalidArgster:NameNotExistent

    The analytics module with the requested name does not exist.

    enc:Receiverter:Actionter:TooManyModules

    There is not enough space in the device to add the analyticsmodules to the configuration.

    env:Receiverter:Action

    ter:ConfigurationConflict

    The device cannot modify the analytics modules without creatinga conflicting configuration.

    env:Senderter:InvalidArgVal

    ter:InvalidModule

    The suggested module configuration is not valid on the device.

    5.3.3.5 DeleteAnalyti cs Modul es

    The following operation deletes multiple analytics modules. If all analytics modules can not bedeleted as requested, the device responds with a fault message.

    Table 10: DeleteAnalyticsModules command

    DeleteAnalyticsModules Access Class: ACTUATE

    Message name Descript ion

    DeleteAnalyticsModulesRequest The request message specifies theVideoAnalyticsConfigurationToken from which the listedAnalytics Modules should be removed.

    tt:ReferenceToken ConfigurationToken [1][1]xs:stringAnalyticsModuleName [1][unbounded]

    DeleteAnalyticsModulesResponse The response is an empty message.

    Fault codes Description

    env:Senderter:InvalidArgster:NoConfig

    The VideoAnalyticsConfiguration does not exist.

    env:Receiverter:Action

    ter:ConfigurationConflict

    The device cannot delete the analytics modules without creatinga conflicting configuration.

    env:Senderter:InvalidArgster:NameNotExistent

    The analytics module with the requested name does not exist.

    5.4 Capabilities

    The capabilities reflect optional functions and functionality of a service. The information isstatic and does not change during device operation. The following capabilites are available:

    RuleSupport: Indication that the device supports rules interface and rules syntax asspecified in Section 5.2

    Analyt icsModuleSu pport:Indication that the device supports the scene analytics moduleinterface as specified in Section 5.3.

  • 7/30/2019 ONVIF VideoAnalytics Service Spec v210

    24/27

    ONVIF 24 Video Analytics. Ver. 2.10

    Table 11: GetServiceCapabilit ies co mmand

    GetServiceCapabilities Access Class: PRE_AUTH

    Message name Description

    GetServiceCapabilitiesRequest

    This message contains a request for device capabilities.

    GetServiceCapabilitiesResponse

    The capability response message contains the requested servicecapabilities using a hierarchical XML capability structure.

    tan:Capabilities Capabilities [1][1]

    Fault codes Description

    No command specific faults!

    5.5 Service speci fic data typ es

    This service does reuse the tt:Config data type defined in the ONVIF Event ServiceSpecification. It does not introduce own data types for the service itself.

    For the data types that describe the scene description refer to the onvif.xsd schema file.

    5.6 Service-specifi c fault codes

    Table 12 below lists the analytics service-specific fault codes. Each command can alsogenerate a generic fault.

    The specific faults are defined as subcode of a generic fault. The parent generic subcode isthe subcode at the top of each row below and the specific fault subcode is at the bottom ofthe cell.

    Table 12: The analytics-specific fault codes

    Parent SubcodeFault Code

    Subcode

    Fault Reason Description

    ter:Actionenv:Receiver

    ter:TooManyRules

    No more spaceavailable.

    There is not enough space inthe device to add the rules tothe configuration.

    ter:Actionenv:Receiver

    ter:TooManyModules

    No more spaceavailable.

    There is not enough space inthe device to add the analyticsmodules to the configuration.

    ter:Actionenv:Receiver

    ter:ConfigurationConflict

    Conflict when usingnew settings

    The new settings result in aninconsistent configuration.

    env:Sender ter:InvalidArgValNo such The requested

  • 7/30/2019 ONVIF VideoAnalytics Service Spec v210

    25/27

    ONVIF 25 Video Analytics. Ver. 2.10

    ter:NoConfig configuration VideoAnalyticsConfigurationdoes not exist.

    ter:InvalidArgValenv:Sender

    ter:InvalidRule

    The rule is invalid.The suggested ruleconfiguration is not valid.

    ter:InvalidArgValenv:Sender

    ter:InvalidModule

    The module is invalid The suggested analyticsmodule configuration is notvalid on the device.

    ter:InvalidArgValenv:Sender

    ter:RuleAlreadyExistent

    The rule exists The same rule name existsalready in the configuration.

    ter:InvalidArgsenv:Sender

    ter:RuleNotExistent

    The rule does notexist

    The rule name or names donot exist.

    ter:InvalidArgsenv:Sender

    ter:NameAlreadyExistent

    The name exists The same analytics modulename exists already in theconfiguration.

    ter:InvalidArgsenv:Sender

    ter:NameNotExistent

    The name does notexist

    The analytics module with therequested name does notexist.

  • 7/30/2019 ONVIF VideoAnalytics Service Spec v210

    26/27

    ONVIF 26 Video Analytics. Ver. 2.10

    Annex A(informative)

    Scene descript ions

    A.1 Co lour Des cr iptor

    A Colour Descriptor is defined as an optional element of the appearance node of an objectnode. The Colour Descriptor is defined by a list of colour clusters, each consisting of a colourvalue, an optional weight and an optional covariance matrix. The Colour Descriptor does notspecify, how the colour clusters are created. They can represent bins of a colour histogram orthe result of a clustering algorithm.

    Colours are represented by three-dimensional vectors. Additionally, the colourspace of eachcolour vector can be specified by a colourspace attribute. If the colourspace attribute is

    missing, the YCbCr colourspace is assumed. It refers to the 'sRGB' gamut with the RGB toYCbCr transformation as of ISO/IEC 10918-1 (Information technology -- Digital compressionand coding of continuous-tone still images: Requirements and guidelines), a.k.a. J PEG. Thecolourspace URI for the YCbCr colourspace is www.onvif.org/ver10/colorspace/YCbCr.

    . . .

    A.2 Class Des criptor

    A Class Descriptor is defined as an optional element of the appearance node of an objectnode. The Class Descriptor is defined by a list of object classes together with a likelihood thatthe corresponding object belongs to this class. The sum of the likelihoods shall NOT exceed 1.

  • 7/30/2019 ONVIF VideoAnalytics Service Spec v210

    27/27

    ONVIF 27 Video Analytics. Ver. 2.10