{
    "version" : "https://jsonfeed.org/version/1",
    "content" : "news",
    "type" : "single",
    "title" : "Five Lessons Learned During a Content Audit |Digital.gov",
    "description": "Five Lessons Learned During a Content Audit",
    "home_page_url" : "/preview/gsa/digitalgov.gov/bc-archive-content-3/","feed_url" : "/preview/gsa/digitalgov.gov/bc-archive-content-3/2016/12/05/five-lessons-learned-during-a-content-audit/index.json","item" : [
    {"title" :"Five Lessons Learned During a Content Audit","summary" : "As part of USAGov’s efforts to provide our audience with the reliable and quality information that they need, this summer, the Health, Education &amp; Benefits (HE&amp;B) topic desk completed its first content audit. Methodology and Results Data informed every step we took. In order to determine which areas to focus on first, the desk gathered","date" : "2016-12-05T12:00:42-04:00","date_modified" : "2025-01-27T19:42:55-05:00","authors" : {"andrea-m-castelluccio" : "Andrea M. Castelluccio"},"topics" : {
        
            "content-strategy" : "Content strategy",
            "product-and-project-management" : "Product and project management"
            },"branch" : "bc-archive-content-3",
      "filename" :"2016-12-05-five-lessons-learned-during-a-content-audit.md",
      
      "filepath" :"news/2016/12/2016-12-05-five-lessons-learned-during-a-content-audit.md",
      "filepathURL" :"https://github.com/GSA/digitalgov.gov/blob/bc-archive-content-3/content/news/2016/12/2016-12-05-five-lessons-learned-during-a-content-audit.md",
      "editpathURL" :"https://github.com/GSA/digitalgov.gov/edit/bc-archive-content-3/content/news/2016/12/2016-12-05-five-lessons-learned-during-a-content-audit.md","slug" : "five-lessons-learned-during-a-content-audit","url" : "/preview/gsa/digitalgov.gov/bc-archive-content-3/2016/12/05/five-lessons-learned-during-a-content-audit/","content" :"\u003cp\u003eAs part of USAGov’s efforts to provide our audience with the reliable and quality information that they need, this summer, the Health, Education \u0026amp; Benefits (HE\u0026amp;B) topic desk completed its first content audit. \u003cdiv class=\"image\"\u003e\n  \u003cimg\n    src=\"https://s3.amazonaws.com/digitalgov/_legacy-img/2017/03/600-x-400-Brainstorming-Brainstorm-Business-People-Design-Concepts-scyther5-iStock-Thinkstock-516755868.jpg\"\n    alt=\"Business people brainstorming around a desk.\"/\u003e\u003c/div\u003e\n\n\u003c/p\u003e\n\u003cp\u003e\u003cstrong\u003eMethodology and Results\u003c/strong\u003e\u003c/p\u003e\n\u003cp\u003eData informed every step we took. In order to determine which areas to focus on first, the desk gathered data from four distinct sources: the Contact Center’s usage of our content to answer customers’ inquiries; site analytics; content inventory and review; and website survey comments. A report and an action plan were produced which would serve as the main driver of our content overhaul.\u003c/p\u003e\n\u003cp\u003e\u003cstrong\u003eSteps We Took and Lessons Learned\u003c/strong\u003e\u003c/p\u003e\n\u003col\u003e\n\u003cli\u003e\u003cstrong\u003eApproach.\u003c/strong\u003e The team decided that all the subtopics were going to be part of the audit, and took a bottom-up analytical approach during the discovery phase. This type of approach occurs when the analysis of the metrics is done without a defined goal. Although we technically had a goal — improve our content — it proved to be too high level. We struggled with the approach because we were trying to find answers when we didn’t know yet the questions to ask.\u003c/li\u003e\n\u003c/ol\u003e\n\u003cpre\u003e\u003ccode\u003e_Lesson learned: Select a representative sample of the content, observe the metrics over time, and clearly define the focus before beginning the work._\n\u003c/code\u003e\u003c/pre\u003e\n\u003col start=\"2\"\u003e\n\u003cli\u003e\u003cstrong\u003eTimeframe for Analysis.\u003c/strong\u003e Time constraints forced us to focus on our objectives and prioritize data that we determined would be most insightful. We arrived at our customer’s page level comments as a good starting point.\u003c/li\u003e\n\u003c/ol\u003e\n\u003cpre\u003e\u003ccode\u003e_Lesson learned: Most things take longer than expected, so more time should be allowed for the discovery portion of a content audit._\n\u003c/code\u003e\u003c/pre\u003e\n\u003col start=\"3\"\u003e\n\u003cli\u003e\u003cstrong\u003eSurvey Comments Analysis.\u003c/strong\u003e What users were telling us in the survey comments was the key to setting our goals. We systemized the analysis in a way that we could determine what would eventually become our content goals: determine what users were trying to do in our pages, how they were trying to do it, and change the negative sentiment of some of our pages with better content. The result of the systematization informed the usability, information architecture tests, and the content analysis heuristics.\u003c/li\u003e\n\u003c/ol\u003e\n\u003cpre\u003e\u003ccode\u003e_Lesson learned: Challenge everyone’s assumptions about the content and users’ interaction with it, and turn survey comments into quantitative data. It helps to easily identify trends and opens up a conversation on user versus “creator.”_\n\u003c/code\u003e\u003c/pre\u003e\n\u003col start=\"4\"\u003e\n\u003cli\u003e\u003cstrong\u003eContent Analysis.\u003c/strong\u003e As part of the discovery phase, there was a content analysis heuristics where we compared existing content with the survey comments and the analytics data, and we made a list of potential content to test, improve, move, or delete. The content analysis heuristics took the form of regular face-to-face meetings of the desk members with a UX team member. We considered issues with the quality of our content, such as usefulness and relevance, clarity and accuracy, completeness, style; with the information architecture such as usability and findability, completeness, consistency, and appropriate structure of our assets.\u003c/li\u003e\n\u003c/ol\u003e\n\u003cpre\u003e\u003ccode\u003e_Lesson learned: We don’t know our web content until we audit it, and all team members are figuring it out together._\n\u003c/code\u003e\u003c/pre\u003e\n\u003col start=\"5\"\u003e\n\u003cli\u003e\u003cstrong\u003eTesting.\u003c/strong\u003e We know that optimization is an ongoing, iterative process, and that next time our content is ready for another audit our objectives and planning will be different.\u003c/li\u003e\n\u003c/ol\u003e\n\u003cpre\u003e\u003ccode\u003e_Lesson learned: There is no such a thing as a standard content audit process. Data (or external circumstances) will set the stage for the next steps. It is imperative to lay out a framework and context for what you’re auditing, why, and what your main objectives are. The goal is always the same — improve the public’s experience with your content._\n\u003c/code\u003e\u003c/pre\u003e\n\u003cp\u003e\u003cstrong class=\"markup--strong markup--p-strong\"\u003eNext Steps\u003c/strong\u003e\u003c/p\u003e\n\u003cp\u003eWith the results of the content analysis and audit in combination with the report of findings and recommendations, we’re revamping our content on various USA.gov pages.\u003cem\u003eAndrea M. Castelluccio, Ph.D., is a bilingual member of the USAGov content team.\u003c/em\u003e\n\u003cem\u003eThis was originally published on the \u003ca href=\"https://blog.usa.gov/\"\u003eUSAGov blog\u003c/a\u003e{.markup\u0026ndash;anchor.markup\u0026ndash;p-anchor}.\u003c/em\u003e\u003c/p\u003e\n"}
  ]
}
