What problem does QTDashboard solve?
Traditionally, management cannot gain proper control of development, because test data comes from multiple systems for defect tracking, test management and release deployment. This results in improper expectations of software release dates, and untimely delays and ultimately losses in revenue due to lack of timely data.
Designed to provide a holistic view of product quality, engineering deliverables and overall testing activities, QTDashboard delivers a unified view, customizable per team, providing a clear data-driven picture of the entire development process. This Quality Intelligence system visualizes Quality & Test for software releases, builds, features, etc. so that both organizations and engineering departments are able to achieve better visibility on the products’ quality throughout SLDC and in Production environment.
How does QTDashboard work?
QTDashboard pulls and presents near-real-time data of product quality by allowing the creation of a dynamic dashboard with customizable charts.
QTDashboard allows integration with multiple data sources as well as managing unlimited users/platforms as well as releases
QTDashboard can be integrated to any test/defect/orchestrating system with an available API or database access
Dashboards created in QTDashboard can be private or shared across the organization and/or specific user groups, with default views that can be personalized to the user role
QTDashboard is accessible via mobile, tablet and computer
Hosting can be done on cloud or on-premises
QTDashboard supports TV displays while automatically refreshing
How can each team benefit from QTDashboard?
Each team can have their own customized dashboard of essential metrics:
Keeping track of various QA/Test metrics e.g.: Test Coverage, Defect Leakage, High Risk Areas, Defect Aging, Defect Validity, Automation result trends, Automation false positive tendency, etc.
Making test strategy dynamic and improving teams’ and products’ quality
Reflecting unified reports of manual and automation tests, and providing automated indications whenever SLA breaches (critical bugs, % of failing automated test cases, insufficient test coverage)
Correlating between areas of delivery and test coverage areas
Keeping track of deliverable KPIs and SLAs (points delivered over time/releases, # of deployments over time, MTTR, etc.)
Providing automated indications when failing to meet delivery-related SLA
Having a clear understanding of the scope of the release and the set of actions needed for the deployment (overall size of the release, any pending issues which might have been missed, required database migrations, release/micro-services dependencies, etc.)
Having a clear understanding of products’/releases’ state in each and every given time
Evaluating overall engineering teams’ performance
Constantly assessing the balance between speed of Delivery vs. Quality reflecting teams’ goals across the entire organization