{
    "version" : "https://jsonfeed.org/version/1",
    "content" : "news",
    "type" : "single",
    "title" : "Making Customer Satisfaction Data Actionable at the Peace Corps |Digital.gov",
    "description": "Making Customer Satisfaction Data Actionable at the Peace Corps",
    "home_page_url" : "/preview/gsa/digitalgov.gov/cm-topics-button-component/","feed_url" : "/preview/gsa/digitalgov.gov/cm-topics-button-component/2018/01/30/making-customer-satisfaction-data-actionable-at-peace-corps/index.json","item" : [
    {"title" :"Making Customer Satisfaction Data Actionable at the Peace Corps","summary" : "Learn how measuring customer experience helped inform and guide website development at Peace Corps.","date" : "2018-01-30T09:30:00-04:00","date_modified" : "2024-04-02T09:45:13-04:00","authors" : {"chris-rottler" : "Chris Rottler"},"topics" : {
        
            "analytics" : "Analytics",
            "content-strategy" : "Content Strategy",
            "customer-experience" : "Customer experience",
            "design" : "Design",
            "research" : "Research"
            },"featured_image" : { "uid" :
  "peace-corps", "alt" :
  "Peace Corps Logo" },"branch" : "cm-topics-button-component",
      "filename" :"2018-01-16-making-customer-satisfaction-data-actionable-at-peace-corps.md",
      
      "filepath" :"news/2018/01/2018-01-16-making-customer-satisfaction-data-actionable-at-peace-corps.md",
      "filepathURL" :"https://github.com/GSA/digitalgov.gov/blob/cm-topics-button-component/content/news/2018/01/2018-01-16-making-customer-satisfaction-data-actionable-at-peace-corps.md",
      "editpathURL" :"https://github.com/GSA/digitalgov.gov/edit/cm-topics-button-component/content/news/2018/01/2018-01-16-making-customer-satisfaction-data-actionable-at-peace-corps.md","slug" : "making-customer-satisfaction-data-actionable-at-peace-corps","url" : "/preview/gsa/digitalgov.gov/cm-topics-button-component/2018/01/30/making-customer-satisfaction-data-actionable-at-peace-corps/","content" :"\n\n\u003cdiv class=\"image image-right\"\u003e\n  \u003cimg\n    src=\"https://s3.amazonaws.com/digitalgov/peace-corps_w200.jpg\" \n    srcset=\"https://s3.amazonaws.com/digitalgov/peace-corps_bu.jpg 48w,https://s3.amazonaws.com/digitalgov/peace-corps_w600.jpg 600w,https://s3.amazonaws.com/digitalgov/peace-corps_w400.jpg 400w,https://s3.amazonaws.com/digitalgov/peace-corps_w200.jpg 200w\"\n    sizes=\"(max-width: 600px) 40vw, 400px\"\n  /\u003e\u003c/div\u003e\n\n\n\u003cp\u003eFor the past couple of years, the \u003ca href=\"https://www.peacecorps.gov/\"\u003ePeace Corps\u003c/a\u003e has used online-based intercept surveys on peacecorps.gov to measure user satisfaction. The data captured over time has been interesting, but has not varied much month-to-month, which has made it difficult to translate insight into actionable enhancements on the website.\u003c/p\u003e\n\u003cp\u003eIn order to get more out of the user satisfaction data, we developed a framework that applies statistical models to the data collected that identify key performance indicators (KPIs) that have the greatest likelihood to increase overall user satisfaction. We hope that this framework will help our team prioritize website changes and enhancements that have the greatest return on investment.\u003c/p\u003e\n\u003cp\u003eIn June 2016, the Peace Corps launched a redesigned website in order to improve user experience and enhance our storytelling capabilities. As part of the project, we began capturing survey data to measure the impact of the redesign on a handful of KPIs pre- and post-launch. We decided to continue to conduct surveys since the launch to monitor user satisfaction over time.\u003c/p\u003e\n\u003cp\u003eOur main KPIs that are tracked in the survey include overall satisfaction, content comprehension, task completion, likelihood to return to the website, net promoter score, and the primary purpose for the visit (i.e., find volunteer opportunities, learn about the agency, learn about a specific topic, contact an appropriate representative, use a site tool, and other).\u003c/p\u003e\n\u003cp\u003eInitially, the data was very interesting. We saw significant improvements to the user experience after the redesign was launched. However, while the initial improvement pre- and post-launch was striking, the delta month-to-month thereafter was minimal.\u003c/p\u003e\n\u003cdiv\n  class=\"image\"\n\u003e\u003cimg\n      src=\"https://s3.amazonaws.com/digitalgov/peace-corps-customer-satisfaction_w800.png\"\n      \n        alt=\"A bar graph showing User Satisfaction: Task Completion vs. Content Comprehension\"\n        srcset=\"https://s3.amazonaws.com/digitalgov/peace-corps-customer-satisfaction_bu.jpg 48w,https://s3.amazonaws.com/digitalgov/peace-corps-customer-satisfaction_w600.png 600w,https://s3.amazonaws.com/digitalgov/peace-corps-customer-satisfaction_w400.png 400w,https://s3.amazonaws.com/digitalgov/peace-corps-customer-satisfaction_w200.png 200w\"\n\n      sizes=\"(max-width: 800px) 100vw, 800px\"\n    /\u003e\u003c/div\u003e\n\n\n\u003cp\u003eThe data provided a good moving indicator of our KPIs, but we found it difficult to choose what area to focus on to increase user satisfaction, and which enhancements to prioritize. To that end, we built a framework that systematizes model development, allowing us to build statistical models against the survey data to measure which KPI has the greatest likelihood to increase satisfaction. In the example above, “User Satisfaction: Task Completion vs. Content Comprehension,” we wanted to understand what KPI – comprehension of content or ability to complete a task – had the most potential to increase overall user satisfaction. As part of the analysis, we segmented the data set by users’ purpose of visit and developed a separate model for each segment to see if content comprehension or users’ ability to complete a task had a greater likelihood to increase overall user satisfaction.\u003c/p\u003e\n\u003cp\u003eBased on the type of data we capture – ordinal in nature – we run \u003ca href=\"https://en.wikipedia.org/wiki/Ordered_logit\"\u003eordered logistic regression\u003c/a\u003e. The chart above is a summary of the statistical outputs. What it means is that for a one unit increase in users’ ability to complete a task or comprehension of content increases the likelihood that their overall satisfaction increases (i.e., by the numbers to the right of each bar). So, for example, the overall satisfaction of users who come to the site to “learn about a specific topic” are 17.47 times more likely to increase to a higher score if the content is easy to understand. The satisfaction score of those same visitors is only 5.11 times more likely to increase if their ability to complete a task increases one unit. This example is fairly basic, but it is meant to illustrate the utility of this approach in terms of prioritizing website enhancements and testing them against customer satisfaction.\u003c/p\u003e\n\u003cp\u003eWe plan to dig deeper by developing unique surveys for specific sections of the site, and developing models against different KPIs and dependent variables that are important to the Peace Corps in addition to satisfaction. Hopefully, this will help us improve the user experience and continue to optimize peacecorps.gov so that it better serves our visitors.\u003c/p\u003e\n"}
  ]
}
