Module 13 - T202
T202: Overview of Test Design Specifications Test Cases, and Test Procedures
HTML of the Student Supplement
(Note: This document has been converted from the Student Supplement to 508-compliant HTML. The formatting has been adjusted for 508 compliance, but all the original text content is included, plus additional text descriptions for the images, photos and/or diagrams have been provided below.)
(Extended Text Description: Large graphic cover page with dark blue background with the title in white letters “T202 - Overview of Test Design Specifications Test Cases, and Test Procedures.” At the middle left is the “Standards ITS Training” logo with a white box and letters in blue and green. The words “Student Supplement” and “RITA Intelligent Transportation Systems Joint Program Office” in white lettering are directly underneath the logo. Three light blue lines move diagonally across the middle of the blue background.)
T202: Overview of Test Design Specifications, Test Cases, and Test Procedures
Table of Contents
System Engineering “V” Diagram - 2
Sample Test Design Specification – DMS - 4
Sample Test Case Specification – DMS - 6
Sample Test Procedure Specification – DMS - 9
Sample Test Case Specification – TSS - 10
Sample Test Procedure Specification – TSS - 12
Sample Test Case Specification – CCTV - 13
Sample Test Procedure Specification – CCTV - 16
Terminology/Glossary - 19
References - 22
Systems Engineering “V” Diagram
The “V” diagram used by the ITS industry, as shown in Figure 1, reflects a customization of the more general systems engineering process (SEP), but follows widely accepted guidelines.
(Extended Text Description: Slide 10: A graphic of the FHWA “V” diagram of the Project Life Cycle is shown. The main graphic is a V-shaped diagram with some additional horizontal “wings” on the left and right side of the top of the V. Starting from the left “wing” the steps are Regional Architecture(s) and Feasibility Study / Concept Exploration. At this point the steps begin to descend the left side of the V with: Concept of Operations; System Requirements; High-Level Design; and Detailed Design. The base of the V shows Software / Hardware Development & Field Installation. At this point the steps begin to ascend the right side of the V with: Unit / Device Testing; Subsystem Verification; System Verification & Deployment; and System Validation. Finally, the right “wing” includes steps that occur past the testing phases: Operations and Maintenance; Changes and Upgrades; and Retirement / Replacement. Each phase on the left of the “V” is associated by means of a dashed arrow with a phase on the right “V” (from the bottom): Detailed Design is associated with Unit / Device Testing by the Unit / Device Test Plan; High-Level Design is associated with Subsystem Verification by the Subsystem Verification Plan (Subsystem Acceptance); System Requirements is associated with System Verification & Deployment by the System Verification Plan (System Acceptance); and Concept of Operations is associated with System Validation by the System Validation Plan. The left side of the “V” is labeled “Decomposition and Definition”, the bottom is labeled “Implementation”, and the right side of the “V” is labeled “Integration and Recomposition”. At the bottom, a rightward facing arrow is labeled “Time Line".)
Figure 1: Systems Engineering “V” Diagram for ITS
The “V” starts with identifying the portion of the regional ITS architecture that is related to the project. Other artifacts of the planning and programming processes that are relevant to the project are collected and used as a starting point for project development. This is the first step in defining an ITS project.
Next, the agency develops a business case for the project. Technical, economic, and political feasibility is assessed; benefits and costs are estimated; and key risks are identified. Alternative concepts for meeting the project's purpose and need are explored, and the superior concept is selected and justified using trade study techniques. Once funding is available, the project starts with the next step.
The “V” then begins to dip downwards to indicate an increased level of detail. The project stakeholders reach a shared understanding of the system to be developed and how it will be operated and maintained. The Concept of Operations is documented to provide a foundation for more detailed analyses that will follow. It will be the basis for the system requirements that are developed in the next step.
At the system requirements stage, the stakeholder needs identified in the Concept of Operations are reviewed, analyzed, and transformed into verifiable requirements that define what the system will do but not how the system will do it. Working closely with stakeholders, the requirements are elicited, analyzed, validated, documented, and version controlled.
The system design is created based on the system requirements including a high-level design that defines the overall framework for the system. Subsystems of the system are identified and decomposed further into components. Requirements are allocated to the system components, and interfaces are specified in detail. Detailed specifications are created for the hardware and software components to be developed, and final product selections are made for off-the-shelf components.
Hardware and software solutions are created for the components identified in the system design. Part of the solution may require custom hardware and/or software development, and part may be implemented with off-the-shelf items, modified as needed to meet the design specifications. The components are tested and delivered ready for integration and installation.
Once the software and hardware components have been developed, the “V” changes direction to an upward slope. The various components are individually verified and then integrated to produce higher-level assemblies or subsystems. These assemblies are also individually verified before being integrated with others to produce yet larger assemblies, until the complete system has been integrated and verified.
Eventually, the system is installed in the operational environment and transferred from the project development team to the organization that will own and operate it. The transfer also includes support equipment, documentation, operator training, and other enabling products that support ongoing system operation and maintenance. Acceptance tests are conducted to confirm that the system performs as intended in the operational environment. A transition period and warranty ease the transition to full system operation.
After the ITS system has passed system verification and is installed in the operational environment, the system owner/operator, whether the state DOT, a regional agency, or another entity, runs its own set of tests to make sure (i.e., validate) that the deployed system meets the original needs identified in the Concept of Operations.
Once the customer has approved the ITS system, the system operates in its typical steady state. System maintenance is routinely performed and performance measures are monitored. As issues, suggested improvements, and technology upgrades are identified, they are documented, considered for addition to the system baseline, and incorporated as funds become available. An abbreviated version of the systems engineering process is used to evaluate and implement each change. This occurs for each change or upgrade until the ITS system reaches the end of its operational life.
Finally, operation of the ITS system is periodically assessed to determine its efficiency. If the cost to operate and maintain the system exceeds the cost to develop a new ITS system, the existing system becomes a candidate for replacement. A system retirement plan will be generated to retire the existing system gracefully.
To find more information on the systems engineering process, see the links provided in the References section of this supplement.
Sample
Test Design Specification - DMS
Test Design Specification Identifier
TDS-C-DMS
Introduction
This Test Design Specification is to clarify and expand on
the details given in Test Plan TP-C-DMS2-1, which defines testing to be
performed on a DMS to verify conformance with NTCIP 1203 v03.
Features to be Tested
All mandatory requirements of NTCIP 1203 v03 shall be tested.
In addition, optional requirements selected in the NTCIP 1203 Protocol
Requirements List (PRL) as shown below shall be tested.
Protocol Requirements List (excerpts)
NTCIP 1203 Protocol Requirements List (PRL)
USER NEED SECTION NUMBER |
USER NEED |
FR SECTION NUMBER |
FUNCTIONAL REQUIREMENT |
CONFORMANCE |
SUPPORT/ PROJECT REQUIREMENT |
ADDITIONAL PROJECT REQUIREMENTS |
---|---|---|---|---|---|---|
2.3.2.3 |
DMS Display Matrix Configuration |
M |
Yes |
The DMS shall be 9,000 millimeters wide (0..65535) and 2,700 millimeters high (0..65535), inclusive of borders. The Sign's Border shall be at least 400 millimeters wide (0..65535) and 400 millimeters high (0..65535). |
||
2.3.2.3.1 |
Non-Matrix |
0.2 (1) |
Yes/No |
|||
2.3.2.3.2 (Matrix) |
Matrix |
0.2 (1) |
Yes/No |
The pitch between pixels shall be at least 66 millimeters (0..255). |
||
2.3.2.3.2.1 |
Full Matrix |
0.3 (1) |
Yes/No |
The sign shall be pixels wide (0..65535) and pixels high (0..65535). |
||
2.3.2.3.2.2 |
Line Matrix |
0.3 (1) |
Yes/No |
The sign shall have lines with each line being pixels wide and pixels high. |
||
2.3.2.3.2.3 |
Character Matrix |
0.3(1) |
Yes/No |
The sign shall be 18 characters wide and 3 characters high with each character being 5 pixels wide (0..255), 7 pixels high (0..255). |
(Additional Notes from the Author: Below the text, an excerpt from the NTCIP 1203 Protocol Requirements List (PRL) shows a table with the following columns: user Need Section Number, User Need, FR Section Number, Functional Requirement, Conformance, Support / Project Requirement and Additional Project Requirements. Entries for User Need 2.3.2.3 through 2.3.2.3.2.3 are shown. Under the Support / Project Requirement column, individual entries indicating “Yes/No” have had either Yes or No highlighted.)
NTCIP 1203 Protocol Requirements List (PRL)
USER NEED SECTION NUMBER |
USER NEED |
FR SECTION NUMBER |
FUNCTIONAL REQUIREMENT |
CONFORMANCE |
SUPPORT/ PROJECT REQUIREMENT |
---|---|---|---|---|---|
2.5.1.2 |
Determine Sign Display Capabilities |
O |
Yes/No |
||
3.5.1.2.1.1 |
Determine the Size of the Sign Face |
M |
Yes |
||
3.5.1.2.1.2 |
Determine the Size of the Sign Border |
M |
Yes |
||
3.5.1.2.1.3 |
Determine Beacon Type |
M |
Yes |
||
3.5.1.2.1.4 |
Determine Sign Access and Legend |
M |
Yes |
||
3.5.1.2.2.1 |
Determine Sign Face Size in Pixels |
Matrix:M |
Yes/NA |
||
3.5.1.2.2.2 |
Determine Character Size in Pixels |
Matrix:M |
Yes/NA |
||
3.5.1.2.2.3 |
Determine Pixel Spacing |
Matrix:M |
Yes/NA |
(Additional Notes from the Author: Above an excerpt from the NTCIP 1203 PRL shows a table with the same column headings as described in slide 32. The Entry for User Need Section Number 2.5.1.2, with FR Section Numbers 3.2.1.2.1.1 through 3.5.1.2.2.3, is shown. The first entry shows User Need Section Number 2.5.1.2, User Need “Determine Sign Display Capabilities”, Conformance “O”, and Support / Project Requirement “Yes / No”, with Yes highlighted. For the three requirements at the bottom of the table the Conformance column indicates “Matrix: M”, and the Support / Project Requirement column shows “Yes / NA”, with Yes highlighted.)
Approach Refinements
The testing approach is to be as defined in the Test Plan
TP-C-DMS2-1. In addition to the use of NTESTER v2.0 as a testing tool,
Intelligent Devices’ DeviceTester shall also be allowed.
Test Identification
Test Cases and Test Case Identifiers are as defined in the
Test Case Specification TCS-C-DMS.
SAMPLE Test Case Specification - DMS
Test Case Specification Identifier
TCS-C-DMS
Purpose
This document defines the test cases to be used to verify
conformance of the DMS with NTCIP 1203 v03. They specify the input and output
parameters and special conditions for use with the published NTCIP 1203 v03
Test Cases/Test Procedures.
Test Items
The items selected for testing are in accordance with the
project Protocol Requirements List (PRL) as given in Test Design Specification
TDS-C-DMS. The required test cases listed in Appendix A were determined by
means of the Requirements to Test Cases Traceability Matrix found in NTCIP 1203
v03.
Input Specifications
The individual input specifications for each test case are
given in Appendix A.
Output Specifications
The individual output specifications for each test case are defined
in the NTCIP 1203 v03 Test Procedures.
Environmental Needs
In addition to the testing facilities defined in the Test
Plan TP-C-DMS2-1, the following Test Cases will require special tools:
Test Case ID |
Special Items Required: |
---|---|
TC 1.2 |
Tape measure, 30m, marked in mm |
TC 1.3 |
Tape measure, 30m, marked in mm |
TC 1.8 |
Tape measure, 30m, marked in mm |
Special Procedural Requirements
No special procedural requirements exist.
Intercase Dependencies
Intercase Dependencies are as detailed in NTCIP 1203 v03.
Appendix A (excerpt)
* NOTE:
Variables with the same name must be assigned the same value. The user only
needs to define variable values for those test cases that are selected.
NTCIP 1203 Test Case/Test Procedure ID |
Test Case/Test Procedure |
Selected |
Variable |
Reference |
Value |
||||||
---|---|---|---|---|---|---|---|---|---|---|---|
1.1 |
Determine Sign Type and Technology |
X |
|
||||||||
1.2 |
Determine the Size of the Sign Face |
X |
|
||||||||
1.3 |
Determine Size of the Sign Border |
X |
|
||||||||
1.4 |
Determine Beacon Type |
X |
|
||||||||
1.5 |
Determine Sign Access and Legend |
X |
|
||||||||
1.6 |
Determine Sign Face Size in Pixels |
X |
|
||||||||
1.7 |
Determine Character Size in Pixels |
X |
|
||||||||
1.8 |
Determine Pixel Spacing |
X |
|
SAMPLE Test Procedure Specification – DMS
The Test Case/Test Procedure below are excerpted from NTCIP 1203 v03, which includes Test Cases and Test Procedures.
Test Case: |
Title: |
Determine Sign Face Size in Pixels |
||||||||||||||||||||||||||||||||||||||||||
Description: |
This test case verifies that the DMS indicates that it has a height and width in pixels that meet the requirements of the specifications. |
|||||||||||||||||||||||||||||||||||||||||||
Variables: |
|
|||||||||||||||||||||||||||||||||||||||||||
Pass/Fail Criteria: |
The DUT shall pass every verification step included within the Test Case to pass the Test Case. |
|||||||||||||||||||||||||||||||||||||||||||
|
||||||||||||||||||||||||||||||||||||||||||||
Test Case Results |
||||||||||||||||||||||||||||||||||||||||||||
Tested By: |
Date Tested: |
Pass / Fail |
||||||||||||||||||||||||||||||||||||||||||
Test Case Notes: |
||||||||||||||||||||||||||||||||||||||||||||
SAMPLE Test Case Specification – TSS
The Test Case Specification below was created for this webinar, as NTCIP 1209, TSS, does not include Test Cases or Test Procedures.
Test Case Specification Identifier
TCS-1209
Purpose
These test cases are to be used to verify conformance of the
TSS (Transportation Sensor System) with NTCIP 1209 v02. They specify the input
and output parameters for use with the Test Procedures TPS-1209.
Test Items
These test cases will provide details of the tests to be
performed to verify conformance of the TSS with NTCIP 1205 v01. The Test Item
is to consist of a Transportation Sensor System. The Test Procedure steps that
support each Test Case are mapped in the table in Appendix A.
Input Specifications
The individual input specifications for each test case are
given in Appendix A.
Output Specifications
The individual output specifications for each test case are
given in Appendix A.
Environmental Needs
No additional needs exist beyond those defined in the Test
Plan.
Special Procedural Requirements
No special procedural requirements exist.
Intercase Dependencies
No intercase dependencies exist.
Appendix A:
Test Case Number |
Test Case Name |
Test Procedure |
Object(s) Under Test |
Variables |
Result |
---|---|---|---|---|---|
TC1205-001 |
Restart TSS |
TP1205-001 |
sensorSystemStatus.0, sensorSystemReset.0 |
Command = ‘restart’ |
sensorSystemStatus.0 = oK |
TC1205-002 |
Reinitialize TSS |
TP1205-001 |
sensorSystemStatus.0, sensorSystemReset.0 |
Command = ‘reinitializeUserSettings’ |
sensorSystemStatus.0 = oK |
TC1205-003 |
Restore TSS |
TP1205-001 |
sensorSystemStatus.0, sensorSystemReset.0 |
Command = ‘restoreFactoryDefaults’ |
sensorSystemStatus.0 = oK |
TC1205-004 |
Retune TSS |
TP1205-001 |
sensorSystemStatus.0, sensorSystemReset.0 |
Command = ‘retune’ |
sensorSystemStatus.0 = oK |
TC1205-005 |
Re-sync TSS |
TP1205-001 |
sensorSystemStatus.0, sensorSystemReset.0 |
Command = ‘resyncSamplingPeriods’ |
sensorSystemStatus.0 = oK |
TC1205-006 |
Run Short Diagnostics |
TP1205-001 |
sensorSystemStatus.0, sensorSystemReset.0 |
Command = ‘shortDiagnostics’ |
sensorSystemStatus.0 = oK |
TC1205-007 |
Run Long Diagnostics |
TP1205-001 |
sensorSystemStatus.0, sensorSystemReset.0 |
Command = ‘fullDiagnostics’ |
sensorSystemStatus.0 = oK |
SAMPLE Test Procedure Specification - TSS
The Test Procedure below was created for this webinar, as NTCIP 1209, TSS, does not include Test Cases or Test Procedures.
Test Procedure: TP1209-001 |
Title: Description: Pass/Fail Criteria: Variables: |
Reset the TSS This Test Case verifies that the operator can correctly reset the TSS. The DUT shall pass every verification step included within the Test Case in order to pass the Test Case. Command Command to be implemented |
|
Test Step Number |
Test Step |
Results |
|
---|---|---|---|
1. |
GET sensorSystemStatus.0 |
Pass / Fail |
|
2. |
IF the RESPONSE VALUE to sensorSystemStatus.0 is ‘initializing’, ‘pendingConfigurationChange’, or ‘validatingPendingConfiguration’ EXIT the Test Procedure, and correct the deficiency before restarting the test. |
||
3. |
SET sensorSystemReset.0 to Command. |
Pass / Fail |
|
4. |
GET sensorSystemStatus.0 |
Pass / Fail |
|
5. |
IF the RESPONSE VALUE for sensorSystemStatus.0 equals 'initializing', then GOTO Step 4. NOTE--If the RESPONSE VALUE remains at ‘initializing’ for more than the maximum TSS initialization time, this test fails. |
Pass / Fail |
|
6. |
VERIFY that the RESPONSE VALUE for sensorSystemStatus.0 equals ‘oK’. |
Pass / Fail |
|
Test Procedure Result: |
Pass / Fail |
||
Test Procedure Notes: |
|||
SAMPLE Test Case Specification – CCTV
The Test Case Specification below was created for this webinar, as NTCIP 1205, CCTV, does not include Test Cases or Test Procedures.
Test Case Specification Identifier
TCS-1205
Purpose
These test cases are to be used to verify conformance of the
CCTV (closed circuit TV) camera with NTCIP 1205 v01. They specify the input and
output parameters for use with the NTCIP 1205 v01 Test Procedures.
Test Items
These test cases will provide details of the tests to be
performed to verify conformance of the camera with NTCIP 1205 v01. The items
selected for testing are in accordance with the Test Design Specifications
TDS-1205. The Test Procedure steps that support each Test Case are mapped in
the table in Appendix A.
Input Specifications
The individual input specifications for each test case are
given in Appendix A.
Output Specifications
The individual output specifications for each test case are
given in Appendix A.
Environmental Needs
No additional needs exist beyond those defined in the Test
Plan.
Special Procedural Requirements
No special procedural requirements exist.
Appendix A:
Test Case Number |
Test Case Name |
Test Procedure Steps |
Object(s) Under Test |
Variables |
Result |
---|---|---|---|---|---|
TC1205-001 |
Zoom absolute - maximum telephoto |
TP1205-003 Steps 4-10 |
positionZoomLens.0 |
Mode = 2 (absolute) Speed_Tele_Fast = 127 Offset = rangeZoomLimit.0 |
Camera should zoom to its maximum setting at its maximum speed |
TC1205-002 |
Zoom absolute - maximum wide |
TP1205-003 Steps 11-17 |
positionZoomLens.0 |
Mode = 2 (absolute) Speed_Wide_Fast = -128 Offset = 0 |
Camera should zoom to its widest angle setting at its maximum speed |
TC1205-003 |
Zoom absolute -half-way |
TP1205-003 Steps 18-24 |
positionZoomLens.0 |
Mode = 2 (absolute) Speed_Tele_Fast = 127 Offset = (rangeZoomLimit.0)/2 |
Camera should zoom halfway to its maximum setting at its maximum speed |
TC1205-004 |
Zoom delta |
TP1205-003 Steps 25-30 |
positionZoomLens.0 |
Mode = 1 (delta) Speed_Tele_Mid = 50 Offset = (rangeZoomLimit.0)/10 |
Camera should zoom 1/10th toward its maximum zoom at midspeed with each delta command |
TC1205-005 |
Zoom continuous |
TP1205-003 Steps 31-33 |
positionZoomLens.0 |
Mode = 3 (continuous) Speed_Wide_Slow = -30 Offset = 0 |
Camera should zoom wide at a slow speed |
TC1205-006 |
Zoom stop |
TP1205-003 Steps 34-38 |
positionZoomLens.0 |
Mode = 0 (stop) Speed = 0 Offset = 0 |
Camera should stop |
TC1205-007 |
Zoom past rangeZoomLimit.0 |
TP1205-003 Steps 39-41 |
positionZoomLens.0; rangeZoomLimit.0 |
Mode = 1 (absolute) Speed_Tele_Fast = 127 Offset = 65535 |
Response Error is badValue |
TC1205-008 |
Zoom for greater duration than timeoutZoom.0 |
TP1205-003 Steps 40-46 |
positionZoomLens.0; timeoutZoom.0 |
Mode = 3 (continuous) Speed_Tele_Slow = 10 Offset = rangeZoomLimit.0 |
Camera begins zooming toward its maximum telephoto setting, but stops after Zoom_Timeout seconds |
SAMPLE Test Procedure Specification - CCTV
The Test Procedure below was created for this webinar, as NTCIP 1205, CCTV, does not include Test Cases or Test Procedures.
Test Procedure: TP1205-003 |
Title: Description: Pass/Fail Criteria: Variables: |
Zoom Camera This Test Case verifies that the operator can correctly zoom the camera in and out. The DUT shall pass every verification step included within the Test Case in order to pass the Test Case. Speed_Tele_Fast [85..127] Speed_Wide_Fast [-85..-128] Speed_Tele_Mid [42..84] Speed_Wide_Slow [-1..-41] Speed_Tele_Slow [1..41] Zoom_Timeout [0..65535] (in milliseconds) |
|
Test Step Number |
Test Step |
Results |
|
---|---|---|---|
1. |
Get rangeZoomLimit.0 |
Pass / Fail |
|
2. |
RECORD this integer value and its two-byte hex value as: >> Zoom_Limit >> Zoom_Limit_Hex |
||
3. |
Set timeoutZoom.0 to 0, turning off this feature. |
Pass / Fail |
|
4. |
Set positionZoomLens.0 to Mode: 2 (absolute) Speed: Speed_Tele_Fast Position or Offset: Zoom_Limit |
Pass / Fail |
|
5. |
Delay for 10 seconds |
||
6. |
Verify that the camera zoomed to its absolute telephoto position. |
Pass / Fail |
|
7. |
Get positionZoomLens.0 |
Pass / Fail |
|
8. |
Verify that the Response Value equals Mode: 02 Speed: Speed_Tele_Fast Position or Offset: Zoom_Limit_Hex |
Pass / Fail |
|
9. |
Get positionQueryZoom.0 |
Pass / Fail |
|
10. |
Verify that the Response Value equals Zoom_Limit. A Response Error of noSuchName indicates that positionQueryZoom.0 is unsupported. |
Pass / Fail / Pass Unsupported |
|
11. |
Set positionZoomLens.0 to Mode: 2 (absolute) Speed: Speed_Wide_Fast Position or Offset: 0 |
Pass / Fail |
|
12. |
Delay for 10 seconds |
||
13. |
Verify that the camera zoomed to its absolute wide position. |
Pass / Fail |
|
14. |
Get positionZoomLens.0 |
Pass / Fail |
|
15. |
Verify that the Response Value equals Mode: 02 Speed: Speed_Wide_Fast Position or Offset: 00 00 |
Pass / Fail |
|
16. |
Get positionQueryZoom.0 |
Pass / Fail |
|
17. |
Verify that the Response Value equals 0. A Response Error of noSuchName indicates that positionQueryZoom.0 is unsupported. |
Pass / Fail / Pass Unsupported |
|
18. |
Set positionZoomLens.0 to Mode: 2 (absolute) Speed: Speed_Wide_Fast Position or Offset: Zoom_Limit / 2 |
Pass / Fail |
|
19. |
Delay for 5 seconds |
||
20. |
Verify that the camera is zoomed in half way |
Pass / Fail |
|
21. |
Get positionZoom.0 |
Pass / Fail |
|
22. |
Verify that the Response Value equals Zoom_Limit_Hex / 2 |
Pass / Fail |
|
23. |
Get positionQueryZoom.0 |
Pass / Fail |
|
24. |
Verify that the Response Value equals Zoom_Limit / 2. A Response Value of 0 indicates that positionQueryZoom.0 is unsupported. |
Pass / Fail / Pass Unsupported |
|
25. |
For each value N, from 1 to 5, perform steps 26 through 30. |
||
26. |
Set positionZoomLens.0 to Mode: 1 (delta) Speed: Speed_Tele_Mid Position or Offset: Zoom_Limit / 10 |
Pass / Fail |
|
27. |
Delay for 2 seconds |
||
28. |
Verify that the camera zooms in by some amount |
Pass / Fail |
|
29. |
Get positionQueryZoom.0 |
Pass / Fail |
|
30. |
Verify that the Response Value has been incremented by Zoom_Limit / 10 from the previous value of this step, where the initial value (N=1) is Zoom_Limit / 2. A Response Error of noSuchName indicates that positionQueryZoom.0 is unsupported. |
Pass / Fail / Pass Unsupported |
|
31. |
Set positionZoomLens.0 to Mode: 3 (continuous) Speed: Speed_Wide_Slow Position or Offset: 0 |
Pass / Fail |
|
32. |
Delay for 5 seconds |
||
33. |
Verify that the camera zooms out continuously |
Pass / Fail |
|
34. |
Set positionZoomLens.0 to Mode: 3 (continuous) Speed: Speed_Tele_Slow Position or Offset: 0 |
Pass / Fail |
|
35. |
Delay for 5 seconds |
||
36. |
Verify that the camera begins zooming in continuously. |
Pass / Fail |
|
37. |
While the camera is still zooming, Set positionZoomLens.0 to Mode: 0 (stopMovement) Speed: 0 Position or Offset: 0 |
Pass / Fail |
|
38. |
Verify that the camera stops zooming. |
Pass / Fail |
|
39. |
IF Zoom_Limit equals 65535, GOTO Step 42 - the camera allows all levels of zoom. |
||
40. |
Set positionZoomLens.0 to Mode: 2 (absolute) Speed: Speed_Tele_Fast Position or Offset: 65535 |
Pass / Fail |
|
41. |
Verify that the RESPONSE ERROR is badValue. |
Pass / Fail |
|
42. |
Set positionZoomLens.0 to Mode: 2 (absolute) Speed: Speed_Wide_Fast Position or Offset: 0 |
Pass / Fail |
|
43. |
Verify that the camera zooms to its widest angle. |
Pass / Fail |
|
44. |
Set timeoutZoom.0 to Zoom_Timeout. |
Pass / Fail |
|
45. |
Set positionZoomLens.0 to Mode: 2 (absolute), Speed: Speed_Tele_Slow Position or Offset: Zoom_Limit |
Pass / Fail |
|
46. |
Verify that the lens begins slowly zooming in, then stops after two seconds. |
Pass / Fail |
|
Test Procedure Result: |
Pass / Fail |
||
Test Procedure Notes: |
|||
Terminology/Glossary
Term |
Definition |
---|---|
Agency Specification |
A document that has been prepared by an agency to define requirements for a subject item or process when procured by the agency. |
Compliance |
A condition that exists when an item meets all of the requirements of an agency specification. |
Concept of Operations |
A document that describes the purpose for a system project, including a description of the current and proposed system, as well as key user needs that the new system is required to address. |
Conformance |
A condition that exists when an item meets all of the mandatory requirements as defined by a standard. It can be measured on the standard as a whole, which means that it meets all mandatory (and applicable conditional) requirements of the standard or on a feature level (i.e., it conforms to feature X as defined in section X.X.X), which means that it meets all mandatory (and applicable conditional) requirements of the feature. |
Device Under Test (DUT) |
NTCIP device that is the object of testing. |
Dialogs |
A sequence of information or message exchanges. |
Environmental Sensor Station (ESS) |
A location that includes a remote processor unit (RPU) connected to one or more sensors for the collection of environmental or meteorological data. It may also include a pavement treatment system (PTS). NOTE-The abbr ESS may also be used in the plural. |
Feature |
A behavior of the device. |
Interchangeable |
A condition which exists when two or more items possess such functional and physical characteristics as to be equivalent in performance and durability, and are capable of being exchanged one for the other without alteration of the items themselves, or adjoining items, except for adjustment, and without selection for fit and performance. NOTE-See National Telecommunications and Information Administration, U.S. Department of Commerce |
Interoperable |
The ability of two or more systems or components to exchange information and use the information that has been exchanged. NOTE-See NTCIP 8004 v02. |
Management Station |
The computer system with which the device communicates. Typically, the management station commands and monitors the device. |
Protocol Requirements List (PRL) |
A table mapping user needs with its associated requirements, along with whether the requirement is mandatory for conformance to the standard. This table allows procurement personnel to specify the desired features of a DMS or can be used by a manufacturer to document the features supported by their implementation. |
Remote Processor Unit (RPU) |
A field processor that collects data from sensors and can communicate the collected data to other computers; the processor may also process the collected data and/or control equipment. |
Requirement |
A condition or capability to which a system shall conform. |
Requirements Traceability |
The ability to follow or study the logical progression among the needs, requirements, and design details in a step-by-step fashion. |
Requirements to Test Cases Traceability Matrix (RTCTM) |
A table that defines the traceability from a requirement to the associated test case. |
Requirements Traceability Matrix (RTM) |
A table that links the requirements to the corresponding dialogs and objects. |
Sensor |
A device that is capable of detecting a condition and reporting the result to an RPU. |
Specification |
A document that specifies, in a complete, precise, verifiable manner, the requirements, design, behavior, or other characteristics of a system or component, and, often, the procedures for determining whether these provisions have been satisfied. |
Systems Engineering Process (SEP) |
Interdisciplinary collaborative approaches to derive, evolve, and verify a life-cycle balanced system solution, which satisfies customer expectations and meets public acceptability. (IEEE) |
Test Case Specification (TCS) |
A document that specifies the actual inputs, predicted results, and set of execution conditions for a test. It also identifies constraints on the test procedures resulting from use of that specific test case. NOTE—See IEEE 829 for a more detailed discussion of test cases. |
Test Design Specification (TDS) |
Per IEEE 829, “A document specifying the details of the test approach for a ... feature or combination of ... features and identifying the associated tests.” For testing NTCIP conformance, this document includes the completed PRL and Requirements to Test Cases Traceability Matrix. |
Test Plan |
A document that prescribes the scope, approach, resources, and schedule of the testing activities. It identifies the items to be tested, the features to be tested, the testing tasks to be performed, the personnel responsible for each task, and the risks associated with the plan. |
Test Procedure Specification (TPS) |
A document that specifies a sequence of actions for the execution of a test. The test procedures test the implementation of the requirement. Test procedures are separated from test design as they are intended to be followed step by step and should not have extraneous detail. |
Validation |
A quality control process that determines if a product fulfills its intended purpose. |
Verification |
A quality control process that determines if a product meets specified requirements. |
References
ITS
IEEE
FHWA