The Logicbroker Blog

Comparative Software Improvement Metric: Objectively scoring two versions of software

Sep 12, 2019, 2:26:03 PM / by Jeff Keltz

Goal

Measuring improvements between two versions of software. 

 

Summary

The Comparative Software Improvement Metric (CSIM or CSI metric) looks to objectively define the potential for improvement and automation in two or more software systems that accomplish the same task. When using the CSI metric, it is important to recognize that it does not look to replace standard practices for improving the user experience (UX) of a system. Rather, it looks to assign a value to the potential (or difference) for automation, or improved backend logic/querying between two similar but different versions of software. 

 

Interaction

The CSIM qualifies all interaction with software in three categories: clicking, typing, and navigating.

  • Clicks are any interaction a user has with their mouse or keyboard where only one “click” is required to move along within the process.
  • Typing is any interaction with fields or inputs that require the user to type something using their keyboard.
  • Navigations (or redirects) are any time the server needs to navigate or bring the user to a different page to either move along the process, or bring the user to the end result. 

 

Due to high variability in the different types of software technologies available, the CSIM does not assess the time involved for a server or computer system to process a request. It assumes that the software design already utilizes the most time-efficient technologies or practices, and only looks to assign a score to evaluate the amount of user interaction required to complete a task. 

 

Scoring

When scoring two versions of software, a lower CSI score assumes a more streamlined process within an application, either through automation, or through improved querying.  

The user (proctor) of the scoring metric assigns a weighted value to each of the interaction types when running the evaluation. Business or system needs determine the ratio for the weighted values. I.e., some users may find manual entry of data to be the biggest slowdown within their system; in this case, the weight for the typing category would hold more value than the weight for the navigation category. When comparing two variants of software, the weighted scale must remain the same between the two evaluations. When displaying results, the weights used to calculate the score must be clearly displayed and an explanation regarding why the weighted values were selected can be provided.

 

Example of Use

Imagine two web applications that allow users to access their internal ID number. The interface for one of the applications (Application-A) may look like this:

The user inputs three fields that are used to query records in the system. They then navigate to the next page, where a dropdown field allows them to select what record they wish to view. Once submitted, they reach the final page containing the desired information: their internal user ID.

**Important notes: Two pages need to load before the user accesses their internal ID. 



Consider a possible re-design: Application-B

Application-B requires only one page load, minimizes user interaction, and contains improved querying that displays data the user may be interested in viewing.

 

While this example may sound extreme (most modern UX designers would never let Application-A get off the ground in that form), it demonstrates the way user interaction is increased if a design that considers streamlining the process is not taken. Some of the key aspects making Application-A less streamlined and usable than Application-B are the number of fields to input (3 vs 1) when making the initial query. Application-A then loads a new page, and requires the user to select which action they want to take. Finally, after selecting the action, clicking submit and then having a new page load, Application-A brings the user to the desired result.

 

In all, if a CSIM score were given to Application-A using an equal weight for all interaction types (1pt), Application-A would garner a score of 8. Here is the calculation:

  • First page: 3 fields to enter, 1 click, 1 page load.
  • Second page: 1 field to click/enter, 1 click, 1 page load.
  • Third Page: Result
  • Total: 8 interactions 

Application-B, on the other hand, would receive a CSIM score of 3. Here is the calculations:

  • First page: 1 entry, 1 click, 1 page load.
  • Second page: Result
  • Total: 3 interactions

 

Application B streamlines the process by allowing the user to enter less fields to achieve the desired result. That is a 62% reduction in the CSIM score and indicates a significant improvement in the design of the software version. 

 

A Real Life Example: Convey Freight Shipments vs Logicbroker + Convey

The Convey freight-scheduling portal allows users to create a bill of lading (BOL) and schedule a delivery pickup time and day for freight deliveries. During this process, users input a number of different fields like:

  • Packaged Items
  • Pallet Descriptions
  • Ship-To Address
  • Ship-From Address
  • Address Type (Residential or Commercial)
  • Package/Pallet Dimensions

 

The UI built within Logicbroker to automate Convey data input minimizes user data entry in the following ways:

  • Automatically selects lowest costing freight service
  • No data entry for ship-to address
  • No packaged Item data entry
  • Destination details pre-populated
  • No data entry for pallet descriptions

 

Samples of the Convey UI/Process

Step 1: Enter all item, dimension, delivery and order information, new page loads

image-36

Step 2: Select desired carrier, new page loads

Step 3: Review information and submit for freight pickup, new page loads. BOL becomes available.

Convey Window-3-edit

After Step 3 is completed, the user clicks submit and then receives the BOL (desired result)

 

Compare that process versus Convey coupled with Logicbroker’s order and freight selection automation, to streamline the process.

 

Step 1: Input shipment data. Order data is pre-populated, shipment is submitted.

Step 2: Logicbroker sends the request to Convey, automatically selects the lowest priced freight provider (backend process), lastly, the BOL is received and populated on the same page.

The following scale was used to calculate the usability improvement between these two software versions:

 

Clicks – 1pt.; Navigation – 1.5 pt.; Typing – 2 pt.

 

Typing received a weight factor of 2 because minimizing the amount a user needs to type greatly reduces the amount of time spent scheduling the freight pickup. As noted in the sample images, Logicbroker uses data received from the order (through an API) to pre-populate a number of fields. 

 

Navigation received a weight factor of 1.5. A primary goal of the project was to have the users input freight information, and then receive the result (final BOL) as quickly as possible, without having to navigate to a different page. This design allows users to receive the BOL without having to work through as many page prompts.

 

Clicking received a weight factor of 1. Due to the nature of each field needing to be clicked before typing, clicking would have an impact on the final score, but is weighed with the smallest relative value because the minimal effect on the overall process. Whenever a field requiring typing was removed, a click was removed as well. That prompted the use of a lower score for click interactions.

 

**Note: Both Logicbroker and Convey provide address book functionality, which reduces the amount of fields that need to be typed for “Ship From” fields. Due to this, a “short use case” and a “long use case” was evaluated for each software version. The evaluation for the long use case did not utilize the address book feature, and the results are displayed below.

Results:

Weights used:

Clicks – 1 pt.; Navigation – 1.5 pt.; Typing – 2 pt.

 

Summary: While Convey’s software requires minimal field input for a freight request form, using Logicbroker in conjunction with Convey allows users to greatly minimize the amount of manual data entry needed (including items, PO number, shipping destination, etc.) These two systems combined provided a total CSIM score improvement of 41% 

 

After completing the CSIM score evaluation, user feedback validated these findings, with some users mentioning an improvement of up to 5 minutes per shipment scheduled. 

 

While the CSIM score does provide an objective scoring metric, it is important to obtain feedback from frequent users of the system to validate all findings. In this way, the data can be thoroughly validated, and any additional areas for improvement may be identified. 

Topics: Blog

Jeff Keltz

Written by Jeff Keltz

Born and raised in Connecticut, but now living in the Washington DC area, Jeff Keltz is an expert regarding data-driven deicisons. His passion for analytics is evident in the way he approaches problems, and how he visualizes solutions. Jeff’s diverse knowledge and understanding includes Computer Programming, Databases, Data Visualization, and Client Delivery. He is a technology-driven people-person, who loves to see brands succeed. Jeff’s expertise is using available technology to be flexible and efficient. Recently, he has helped large and small brands alike go from eCommerce “newbies”, to technology connoisseurs. Always leveraging data in his decisions, Jeff finds ways to assist brands increase their success, understanding, and positive reputation within the market. “Never forget the ‘Three T’s’: Technology, Teamwork, and Tacos” – Jeff Keltz