{
    "version" : "https://jsonfeed.org/version/1",
    "content" : "news",
    "type" : "single",
    "title" : "The Content Corner: Can Automated Content Creation Help Your Agency? |Digital.gov",
    "description": "The Content Corner: Can Automated Content Creation Help Your Agency?",
    "home_page_url" : "/preview/gsa/digitalgov.gov/bc-archive-content-3/","feed_url" : "/preview/gsa/digitalgov.gov/bc-archive-content-3/2015/08/10/the-content-corner-can-automated-content-creation-help-your-agency/index.json","item" : [
    {"title" :"The Content Corner: Can Automated Content Creation Help Your Agency?","summary" : "Nearly half of companies recently surveyed said that automating content creation would save their content marketing teams the most time. We’ve already covered Natural Language Generation (NLG) algorithms and how they have made some forms of automated content generation a reality already, such as for sports recaps or financial data reporting. Let’s take a deeper","date" : "2015-08-10T11:00:23-04:00","date_modified" : "2025-01-27T19:42:55-05:00","authors" : {"tyrus-manuel" : "Tyrus Manuel"},"topics" : {
        
            "content-strategy" : "Content strategy"
            },"branch" : "bc-archive-content-3",
      "filename" :"2015-08-10-the-content-corner-can-automated-content-creation-help-your-agency.md",
      
      "filepath" :"news/2015/08/2015-08-10-the-content-corner-can-automated-content-creation-help-your-agency.md",
      "filepathURL" :"https://github.com/GSA/digitalgov.gov/blob/bc-archive-content-3/content/news/2015/08/2015-08-10-the-content-corner-can-automated-content-creation-help-your-agency.md",
      "editpathURL" :"https://github.com/GSA/digitalgov.gov/edit/bc-archive-content-3/content/news/2015/08/2015-08-10-the-content-corner-can-automated-content-creation-help-your-agency.md","slug" : "the-content-corner-can-automated-content-creation-help-your-agency","url" : "/preview/gsa/digitalgov.gov/bc-archive-content-3/2015/08/10/the-content-corner-can-automated-content-creation-help-your-agency/","content" :"\u003cdiv class=\"image\"\u003e\n  \u003cimg\n    src=\"https://s3.amazonaws.com/digitalgov/_legacy-img/2015/08/600-x-400-Process-of-creating-site-Process-coding-and-programming-Design-and-programming-enotmaks-iStock-Thinkstock-464986388.jpg\"\n    alt=\"Graphic illustration of a machine that processes pieces of content and produces a web page.\"/\u003e\u003c/div\u003e\n\n\n\u003cp\u003eNearly half of companies recently surveyed said that \u003ca href=\"http://www2.getchute.com/l/16442/2015-05-20/\"\u003eautomating content creation\u003c/a\u003e would save their content marketing teams the most time. We’ve already covered \u003ca href=\"/preview/gsa/digitalgov.gov/bc-archive-content-3/2015/08/03/the-content-corner-rise-of-the-machines/\"\u003eNatural Language Generation (NLG) algorithms\u003c/a\u003e and how they have made some forms of automated content generation a reality already, such as for sports recaps or financial data reporting. Let’s take a deeper look at how NLG can help your agency rapidly deploy new content and provide a more personalized content experience for users.\u003c/p\u003e\n\u003ch2 id=\"can-it-help-your-agency\"\u003eCan it Help Your Agency?\u003c/h2\u003e\n\u003cp\u003eThe two main benefits of NLG algorithms are how quickly they can create new content (especially for an urgent item, like an earthquake) and the coverage of more topics without additional labor that they allow. They can also allow humans to spend their time covering more nuanced topics while letting NLG bots handle narratives based upon data analysis.\u003c/p\u003e\n\u003ch3 id=\"push-button-data-narratives\"\u003e\u003cstrong\u003ePush Button Data Narratives\u003c/strong\u003e\u003c/h3\u003e\n\u003cp\u003eOne of the first areas within the federal sphere where NLG algorithms could be leveraged would be to develop narratives based around \u003ca href=\"https://analytics.usa.gov/\"\u003eanalytics.usa.gov\u003c/a\u003e trends. The quality of data generated by that site seems like a perfect place to begin experimenting. In addition to the wonderful raw data displayed on the site, how about an NLG-written blog providing a narrative about the trends over the past week? I had an algorithm automatically analyze the usage data of the main site I support (\u003ca href=\"https://oscar.uscourts.gov/\"\u003eoscar.uscourts.gov\u003c/a\u003e) to produce a quick report for the month of June. I could see something similar being done for analytics.usa.gov. Below is an example, from my report:\u003c/p\u003e\n\u003ch4 id=\"traffic-sources\"\u003eTraffic Sources\u003c/h4\u003e\n\u003cblockquote\u003e\n\u003cp\u003eDirect traffic dropped off to 6,958 sessions from 7,687 sessions, accounting for 41% of your site’s traffic overall. A week earlier, direct traffic made up 43% of all sessions to your site. Organic search was responsible for 52% of your site’s traffic last week with 8,771 sessions, which was in line with the 8,820 sessions from one week before.\u003c/p\u003e\n\u003c/blockquote\u003e\n\u003cp\u003eAlso based on the amount of traffic that analytics.usa.gov reports regarding popular federal weather sites, perhaps data from the National Weather Service and other branches of NOAA could also be included in the NLGs sources to allow for more interpretation of traffic. For example, correlating hurricane or tropical storm activity in the Gulf with a spike in traffic to various NWS sites?\u003c/p\u003e\n\u003cp\u003eUsing NLG for the creation of nearly effortless internal data analysis reports for agencies also seems like a solid starting point for this technology. Instead of sending your manager charts, graphs, and tables of page view data, how about a narrative? And instead of you needing to take hours to analyze the data yourself (and perhaps missing a trend or a connection) all you have to do is push a button.\u003c/p\u003e\n\u003cp\u003eAlong with using NLG to produce standard data analysis reports, you can also use this initial auto-generated product as your starting point and then flesh it out accordingly. For example, you could have an NLG run a \u003ca href=\"https://quillconnect.narrativescience.com/story/d9332a1795d955f89cca3f3a0d0d4497/\"\u003equick analysis of your agency twitter feed\u003c/a\u003e (I ran a quick one for my personal feed), and then use that as the basis for a full report that you later would share as part of your \u003ca href=\"/preview/gsa/digitalgov.gov/bc-archive-content-3/2015/06/29/creating-a-social-media-report-that-matters/\"\u003esocial media report\u003c/a\u003e.\u003c/p\u003e\n\u003ch3 id=\"highly-personalized-content-creation\"\u003e\u003cstrong\u003eHighly Personalized Content Creation\u003c/strong\u003e\u003c/h3\u003e\n\u003cp\u003eAnother area where NLG seems to be in the early stages of its potential is in allowing for the generation of highly personalized content. Despite living in an age where advertising especially has become downright creepy in its personalization levels, the actual content delivered is still not uniquely developed for you. Browsing patterns and history, combined with likes or comments within social media allows for pre-existing content to be served based on preferences. NLG will allow for the generation of content that is specifically made for you based on various data available. One area where this has the potential to save lives is within healthcare.\u003c/p\u003e\n\u003cp\u003eThrough the \u003ca href=\"https://wiki.nci.nih.gov/display/VKC/Open+Health+Natural+Language+Processing+%28OHNLP%29+Consortium\"\u003eOpen Health Natural Language Processing (OHNLP) Consortium\u003c/a\u003e and at least \u003ca href=\"https://wiki.nci.nih.gov/display/VKC/OHNLP+News#OHNLPNews-TwochallengegrantshavebeenfundedthatwillbuilduponcTAKES,MayoClinic%27sClinicalTextAnalysisandKnowledgeExtractionSystem,releasedthroughtheOpenHealthNaturalLanguageProcessingConsortium(OHNLP)inMarch2009.\"\u003etwo grants in 2009\u003c/a\u003e, NIH has been supporting efforts to use NLG to provide the best health information possible to patients in the most accessible format.\u003c/p\u003e\n\u003cp\u003eOne of the provided grants was for \u003ca href=\"http://ctakes.apache.org/\"\u003ecTAKES\u003c/a\u003e, which started at the Mayo Clinic and has since grown and become an Apache incubator project.\u003c/p\u003e\n\u003cp\u003ecTAKES leverages Apache’s UIMA (Unstructured Information Management) applications to scan large volumes of electronic medical records to discover patterns, trends, causal relationships, and temporal events. This data can then be provided to the physician in an understandable, narrative format. As shown in the example below, cTAKES gathers an array of examination information and in the end provides plans and recommendations that are in simple sentences.\u003c/p\u003e\n\u003cp\u003e\u003ca href=\"https://s3.amazonaws.com/digitalgov/_legacy-img/2015/08/1195-x-721-cTAKES-example.jpg\"\u003e\u003cdiv class=\"image\"\u003e\n  \u003cimg\n    src=\"https://s3.amazonaws.com/digitalgov/_legacy-img/2015/08/600-x-362-cTAKES-example.jpg\"\n    alt=\"Apache cTAKES™ is a natural language processing system for extraction of information from electronic medical record clinical free-text.\"/\u003e\u003c/div\u003e\n\n\u003c/a\u003e\u003c/p\u003e\n\u003cp\u003eOther projects such as Migraine, Piglit, OPADE, and HealthDoc are in various stages of trials to provide direct feedback to a patient using NLG to draw from various sources such as user feedback, medical records, or existing sources such as drug databases. All four projects show promise in generating personalized content for migraine sufferers or diabetes patients such as the PIGLIT example shown below.\u003c/p\u003e\n\u003cp\u003e\u003ca href=\"https://s3.amazonaws.com/digitalgov/_legacy-img/2015/08/800-x-530-PIGLITex.jpg\"\u003e\u003cdiv class=\"image\"\u003e\n  \u003cimg\n    src=\"https://s3.amazonaws.com/digitalgov/_legacy-img/2015/08/600-x-398-PIGLITex.jpg\"\n    alt=\"Screen capture of a definition of the drug, Bezafibrate, via the PIGLIT project.\"/\u003e\u003c/div\u003e\n\n\u003c/a\u003e\u003c/p\u003e\n\u003cp\u003eOne other area where NLG seems to be making a big difference (and actually was an early adopter of the technology) is in weather forecasting. As far back as at least 2001, NWS was leveraging natural language and speech generation for accessible weather forecasts, especially in an emergency. Similar efforts have been made in generating text forecasts to quickly provide critical weather information for oil rig operators for instance. A specific effort for a UK system actually found that \u003ca href=\"http://www.sciencedirect.com/science/article/pii/S0004370205000998\"\u003eNLG text was easier to understand than human-generated\u003c/a\u003e due to the analysis done to develop word choice and a need for clarity and consistency.\u003c/p\u003e\n\u003cp\u003eI can also see how NLG could help create better content by programming the optimal headline, sentence, and paragraph lengths and removing common cliches. Sorry, that last sentence made me feel a cold hand of a robot touching me on my shoulder, motioning for me to get out of its way (shivers).\u003c/p\u003e\n\u003ch2 id=\"importance-of-structured-content\"\u003eImportance of Structured Content\u003c/h2\u003e\n\u003cp\u003eA common theme that runs through the use of NLG is its reliance on patterns and structured data. cTAKES is only able to be successful by leveraging UIMA that provides structure to varied electronic medical records. The usage of an established vocabulary is critical as machines are being asked to “understand” health and medical information. One of the ways to foster this understanding is the U.S. National Library of Medicine (part of NIH) and their work with the \u003ca href=\"http://www.nlm.nih.gov/research/umls/about_umls.html\"\u003eUnified Medical Language System (UMLS)\u003c/a\u003e. This system is a complex attempt to link vocabularies, categories, and lexical programs to allow for the interoperability that we strive for within our \u003ca href=\"/preview/gsa/digitalgov.gov/bc-archive-content-3/2014/05/05/government-open-and-structured-content-models-are-here/\"\u003eopen and structured content\u003c/a\u003e efforts throughout this community.\u003c/p\u003e\n\u003cp\u003eNWS also provides similar resources for weather forecasting with its \u003ca href=\"http://graphical.weather.gov/xml/\"\u003eNational Digital Forecast Database XML Web service\u003c/a\u003e. Providing data in a structured format allows users to leverage that data as needed.\u003c/p\u003e\n\u003cp\u003eWhen needing to pull data from a wide variety of sources in order to generate natural language narratives via computer, it is extremely helpful that datasets be predictable. Such as entering geographical data in the same field, being consistent in terms used for states (spell out or abbreviate and how is it abbreviated?), spell out pharmaceuticals, enter weeks, days or months in a consistent format and in the correct fields. Sound familiar? Its essentially all metadata. And NLG depends heavily upon this structure of data (and the structure built around its word and grammar templates) to generate the most readable content possible. Without this attention to detail and structure in the beginning, the push button convenience starts to fall apart. Again, why open and structured content is so important, it allows for yet undiscovered uses of your data to be possible if the structure is there.\u003c/p\u003e\n\u003cp\u003eDespite predictions that over \u003ca href=\"http://www.wired.com/2012/04/can-an-algorithm-write-a-better-news-story-than-a-human-reporter\"\u003e90% of all news will be computer-generated within 15 years\u003c/a\u003e, we should embrace the opportunities that NLG can provide us and remember that it is a tool or an aid, not another machine come to take away our jobs. As we all are faced with doing more with less, being able to auto-generate easy to read reports based upon large amounts of data can be one less thing that we have to expend resources on. To be honest, despite regularly reviewing the lines and lines of analytics data my site generates, I found the NLG-generated report for my site to be revelatory. And more important, easy-to-read for anyone with interest in the site’s data trends and performance. My effort to create a five page report narrative with charts? I had to copy and paste a URL and login to my analytics account (phew, I need a nap).\u003c/p\u003e\n\u003cp\u003eEven if the machine provides only a head-start, that is still a time-savings that allows us to devote our time elsewhere. The main focus should be on delivering the best product possible to our audience and NLGs can help us provide content faster and possibly even cover more diverse or niche subject matter or improve their health and their lives.\u003cem\u003eYou’ve just finished reading the latest article from our Monday column, \u003ca href=\"/preview/gsa/digitalgov.gov/bc-archive-content-3/topics/content-strategy/\"\u003eThe Content Corner\u003c/a\u003e. This column focuses on helping solve the main content issues facing federal digital professionals, including producing enough content and making that content engaging.\u003c/em\u003e\u003c/p\u003e\n"}
  ]
}
