March Meeting: Reporting on the State of DITA

The highlight of last night’s meeting of the PDX DITA User’s Group was Keith Schengili-Roberts’ presentation on The State of DITA 2015. Using the crude metric of “more people to feed at our meetings” over the last few years, we had observed that interest in DITA is growing, but Keith provided a more expansive view of the subject based on his research from the last decade. Here is a sampling of data-based observations drawn from his analysis of job postings, case studies, presentations, LinkedIn references and individual reports:

  • Many hundreds of companies worldwide use DITA, with a concentration in the United States and specifically in California.
  • Computer software might still be the largest individual industry using DITA, but a large array of industries outside software make up the lion’s share of users.
  • DITA is the dominant flavor of XML cited in technical writer job postings (DocBook appears rarely these days)
  • In United States job postings, demand for XML experience is trending up and demand for traditional tech writer tools that don’t require structured authoring is gradually trending down.
  • DITA experience is increasingly required or preferred in job descriptions.
  • Job postings asking for DITA expertise are offering higher starting salaries on average than job postings requiring FrameMaker expertise.

The Q&A covered DITA and aerospace standards, demand in the DITA-based CMS market, how review processes work in a DITA-based documentation organization, and how to get better PDFs out of DITA. And Mark Giffin, who called in from California (he is on the OASIS Lightweight DITA committee), alerted us to the existence of an open-source Markdown-to-DITA plugin. This should be of interest to DITA users who work with programmers who see DITA as an obstacle to collaboration.

Thanks to Keith for providing such a great and engaging talk! You can read more about Keith’s work at his blog, Keith’s presentation was sponsored by Ixiasoft who happen to be his employer as well as a maker of DITA component content management systems. (Thanks to Leah d’Emilio for setting up the tech side and making sure everything went smoothly.)

We were especially pleased to see a handful of new DITA users turning up to explore and network: if you’ve been dithering about coming to a meetup because you’re not yet using DITA, please consider this an invitation to show up and find out more. Also, we discovered proudly that one of our regular attendees found a new job by networking at one of our meetings! We like to be socially useful as well as charming so this was very gratifying news. Maybe you will be next.



March Presentation: Keith Schengili-Roberts on the State of DITA

We’re very excited to be able to feature a remote presentation on “The State of DITA 2015” at our March meeting (see the sidebar on our main page for time, date, and location details). In-person attendees will gather for dinner as usual at 6:30, with the presentation starting at 6:45. If you want to attend remotely, please drop a line to

Keith Schengili-Roberts, DITA Information Architect for IXIASOFT and the writer behind the popular “” blog, has been doing extensive research on who is using DITA, where they are using it, what tools they are using and why. He has surveyed the technical writing marketplace in the United States and the role that DITA skills and experience have come to play in it. If you want to get a better sense as to who is using DITA, what software tools are popular and the many ways in which DITA is being used worldwide, come to this presentation!

Want to Present in March?

We’re accepting submissions for a short presentation at our March 18th meeting, so if you’ve had the stirrings of a DITA topic in your back pocket, please pull it out, brush off the lint, and turn it into a 20-minute talk. We would love to hear from you on any topic related to DITA XML.

To accept the mantle of grandiloquence, just drop a line to before March 1st.

PDX DITA + WritetheDocs Meetup = Magic

A quick and enthusiastic report on last night’s PDX DITA holiday potluck with special guests from the WritetheDocs PDX Meetup Group. But first, a picture of happy documentarians.


We packed 30 people into our largest conference room for a delicious potluck (thanks Puppet team for the buffalo wings!) and a short presentation introducing DITA to prospective users. Leona Campbell and Melanie Jennings enthusiastically described the benefits and challenges of DITA as well as sharing experiences about what it’s like to get up and running with DITA when you’ve previously worked with different toolsets. Because we had a range of experiences at the table, from DITA consultants to working DITA writers to total newbies, there was a great ensuing discussion about why you’d want to use DITA rather than another tool. We also covered the need for different kinds of tooling depending on scale, the challenges of converting existing content versus writing topic-based content in DITA from scratch, and the always popular question of just how challenging it is to teach yourself DITA.

Another great outcome was the robust audience recommendations of resources, both print and online. We’ll be adding to the Resources section of this site soon, so stay tuned.

A big thank you to Mike Jang of WritetheDocs Meetup PDX for the opportunity to join forces, and to Melanie and Leona for a wonderful presentation. We’re looking forward to hearing more from our attendees, especially those who are starting up their DITA pilots soon.



Upcoming Holiday Potluck: WritetheDocs PDX Joins Us!

Our annual Holiday Potluck and final meeting of the year is upcoming on December 3rd. This year we’re featuring a special presentation aimed at new and prospective DITA users, which will also be a joint gathering with members of the WritetheDocs PDX Meetup Group. (We Jivers had a great time attending the WritetheDocs conference this past year, but DITA was not talked about in the presentations. That made us wonder why, and this event is one of the results.)

If you’ve been waiting to show up because you feel like you don’t know where to start learning about DITA, this would be a great time to dip your toes in the water. If you’re a grizzled veteran of DITA and/or PDX DITA, we want to see you too! We’d love to hear about your year in DITA and talk about what you’re doing next.

After last year’s potluck we feel confident in promising a delicious holiday feast. (Please bring a dish! We’re two blocks from Whole Foods and the carts, if you need inspiration.). Then Jive’s Melanie Jennings and Leona Campbell will present a brief introduction to writing in DITA XML, followed by Q&A and general conversation. As usual, we’ll provide beer, questionable witticisms, and stunning views of the Portland skyline.

Looking forward to seeing you!


Integrating the Oxygen WebHelp Plugin into the DITA Open Toolkit

I led a project earlier this year to convert my organization’s legacy Eclipse Help Center builds to HTML5 builds using Oxygen’s new WebHelp plugin for the DITA Open Toolkit. By changing the build output, we achieved the following:

  • Completely eliminated the costly Eclipse server maintenance
  • Improved our analytics results and potentially our SEO
  • Brought our output into HTML5 compliance

Unfortunately, actually implementing the plugin wasn’t just a matter of plopping the plugin into the Plugins directory. I had to upgrade the toolkit, rework scripts, and restructure directories to create a less-confusing, easily upgradeable build. Read on to understand my choices and learn from my experience.

Don’t look here to find a step-by-step procedure for implementing the Oxygen WebHelp plugin. Instead, read this blog post if you want help preparing for this process.

Upgraded the Toolkit

First, I upgraded our toolkit, which I was happy to do because it’s good practice, as well as nice to take advantage of the latest that DITA has to offer. Unfortunately, I couldn’t upgrade it too far, because the plugin only supports DITA-OT 1.7.5, which is still better than the 1.5.4 version we’d been using.

This time, I customized using the Customization directory only. My first experience with DITA was a highly customized DITA OT 4.2.1 that still makes me twitch when I look at it. It worked fine and it had some decent customizations, but upgrading a hacked-up toolkit is about as simple as string theory.

Reworked the Scripts

I reworked the .sh script and our ant scripts. The WebHelp plugin includes a file that you need to edit in order to set environmental variables for the build. This replaces the toolkit’s OOB and was actually a little tricky to figure out because I was retrofitting to an existing build layout.

The instructions from Oxygen tell you how to use the, but don’t mention anything about ant builds, which is how we build here (via Jenkins). Oxygen’s support team was also less familiar with this approach. The reason I didn’t want to use only the was that I didn’t want to have every writer set TRANSTYPE or DITAVAL_FILE or DITA_DIR  as environment variables every time she ran it from the command line.

For example, here’s what it would look like if the writer had to set all of the variables from the command line each time she ran a build:

"/usr/bin/java" -Xmx512m -classpath "docs/DITA-OT1.7.5LBC/tools/ant/lib/ant-launcher.jar" "-Dant.home=docs/DITA-OT1.7.5/tools/ant" -lib "docs/DITA-OT1.7.5LBC/" -lib "docs/DITA-OT1.7.5LBC/lib" -lib "docs/DITA-OT1.7.5/lib/saxonb9-1-0-8j/saxon9.jar" -lib "docs/DITA-OT1.7.5LBC/lib/saxonb9-1-0-8j/saxon9-dom.jar" -lib "docs/DITA-OT1.7.5/plugins/com.oxygenxml.webhelp/lib/license.jar" -lib "docs/DITA-OT1.7.5LBC/plugins/com.oxygenxml.webhelp/lib/log4j.jar" -lib "docs/DITA-OT1.7.5/plugins/com.oxygenxml.webhelp/lib/resolver.jar" -lib "docs/DITA-OT1.7.5plugins/com.oxygenxml.webhelp/lib/ant-contrib-1.0b3.jar" -lib "docs/DITA-OT1.7.5LBC/plugins/com.oxygenxml.webhelp/lib/lucene-analyzers-common-4.0.0.jar" -lib "docs/DITA-OT1.7.5/plugins/com.oxygenxml.webhelp/lib/lucene-core-4.0.0.jar" -lib "docs/DITA-OT1.7.5/plugins/com.oxygenxml.webhelp/lib/xhtml-indexer.jar" -f "docs/DITA-OT1.7.5/build.xml" "-Dtranstype=webhelp" "-Dbasedir=docs/sbs/7_0/trunk/src/dita/" "-Doutput.dir=docs/sbs/7_0/trunk/src/dita/out/webhelp" "-Ddita.temp.dir=docs/sbs/7_0/trunk/src/dita/temp/webhelp" "-Dargs.filter=docs/filters/on_prem_sys_admin_7_0.ditaval" "-Ddita.input.valfile=docs/filters/on_prem_sys_admin_7_0.ditaval" "" "-Ddita.dir=docs/DITA-OT1.7.5" "-Dargs.xhtml.classattr=yes" "-Dargs.input=docs

Instead, I was able to revise the script so the build command now looks like this:

ant -Dargs.filter=docs/filters/on_prem_sys_admin_7_0.ditaval

My Goals for Reworking the Scripts

  • Simplify the build script so more than one user can build from their own environment
  • Automate the build
  • Produce PDFs along with HTML5 output
  • Make it possible to build using numerous maps, filters, and help sets

My Solutions for the Scripts

To reach these goals, I set the variables at build time using a series of build files that are triggered in one simple call on the command line that passes only the DITAVAL_FILE filter. Any writer, or build automation tool, can now use this command because the build scripts are no longer dependent on absolute paths to the DITA-OT or to an SVN repo.

On the command line or in Jenkins, any writer can run the new command in any help set directory, and the local build files reach out to a shared build file for all help sets. I put all of the targets in that one shared file, single-sourcing it like a good writer would.

I also wanted to preserve a previously-used script in our new build. It starts with a simple ant target that calls some .jars for creating PDFs. This target and its supporting .jars comb the main ditamap for submaps and create PDFs for each submap. This Chapter PDF output has been popular with our users, so I made sure we kept it.

Restructured the Directories

Because I was rewriting build files and creating new Jenkins scripts anyway, I figured it was a good time to flatten our directory structure a bit. I removed some unused directories, including some old build cruft and unnecessary branches (all highlighted in red).

Our former structure with extra layers:

  • productA/feature/ (used for eclipse output).
  • productA/version1/trunk/src/dita/
  • productA/version1/trunk/src/build/
  • productA/version1/trunk/output/
  • productA/version1/trunk/src/temp/

Here’s our new simpler way:

  • productA/version1/dita/
  • productA/version1/build/
  • productA/version1/output/
  • productA/version1/temp/

A next step is to research best practices for directory structures and DITA to further simplify our structure, but I’m happy to work in this slightly less cluttered place.

Customized WebHelp CSS and JavaScript

Knowing the hell of a customized toolkit made me extra concerned with customizing the Webhelp plugin. Although the plugin offers a lot, it does not provide a Customization directory. In order to change the look and feel of our HTML5 output, we had to edit the CSS and some JavaScript directly. This means that when/if we upgrade, it will take some work to transfer the changes to the new plugin. I already saw a huge reorg of files and directories between the pre-release and January versions of the same release.

Out of the box, Oxygen WebHelp plugin output looks pretty good:

Oxygen OOB Output

But I wanted to change our help to match our corporate style:
Jive Help Screen

Things to Ponder

  • Customizing the plugin might cause future headaches, but who wants OOTB?
  • Speaking of customizations: If you want to change the banner/header, brace yourself!
  • You will become an expert if you customize this plugin.
  • Upgrading even a non-customized toolkit is not always simple.
  • Oxygen Support is in Romania — an entire workday of waiting for responses is frustrating. But they are helpful and thorough, so ask them lots of questions.
  • Read the Oxygen Support forums ( There are some nuggets out there.
  • Hold out for the most stable version of the plugin, no matter how excited you are about it.
  • The Search and CSS are way better than in any other product we evaluated, but still clunky.
  • My last gripe about customizing this plugin is the utter lack of documentation. The paragraph on customizing the CSS cracked me up because they make it sound like you can use any ol’ CSS file, but you still need to point to the html elements that the plugin uses, such as 5 nested UL elements!

Was It Worth It?

Hell, yeah! I’m known for jumping head first into complicated projects, and this was no exception. (Someday, I’ll blog about improving PDF output using XSLT.) But, should YOU use this plugin? The answer is yes if you:

  • Want HTML5 output for an okay price.
  • Like the out-of-the-box look, or you want to learn (or already know) JavaScript and CSS.
  • Are comfortable using Support forums or Technical Support for help.

Have any of you had to implement the Oxygen WebHelp plugin? If yes, I’d love to hear your challenges and how you overcame them.

Getting Started with Filtering

In my last post about filtering, I tried to explain the theory of filtering, so you’d understand which kinds of problems filtering might solve. In this post, I’ll attempt to show how to get started on an actual project. This post is intended for folks who are just getting started with filtering.

Overall Process for a Filtering Project

Generally speaking, I find my process to be something like this:

  1. Create a matrix of the product’s variations and the associated publications so that I have a reference to consult for my sanity. Basically, I’m trying to sketch out which topics I’ll include and how to filter them in or out (in other words, show them or not show them in the final publication). This might be organized by features, audience, revision number, platform, or a combo of any of these.
  2. Use the matrix to define the values I’m going to need in my filter files. A “value” is the definition behind the features, audience, platform, etc. For example, the values might be “beginner,” “intermediate,” or “advanced,” or “Mac,” “Windows,” or “Linux”. Defining the values is basically figuring out who the audience is and the content they’ll need in the publication.
  3. Create the actual filter files (based on the values in the matrix) so that the computer knows how to show or not show tagged topics and inline text appropriately at build/publish time.
  4. Tag the content in the source files for the different features.
  5. Create the final publications (.doc, html, pdf, etc.) using each filter.

Create a Matrix

In my last post, I used the robot product as an example. The robot came in different models, so the matrix I created for that was based on features. But I find that when I change the example, that sometimes produces a different “aha” moment for readers. So this time around, I’m going to use a cookbook example that’s based on audience values (beginner and advanced).

I need to create a cookbook for two culinary school courses: Beginning Baking and Advanced Baking. I’ve determined the chapters I’m going to need and which ones I’ll be able to share between the two different cookbooks. Here’s how it looks:

Define the Values

So in this case, we’re going to need two audience values:

  • Beginner
  • Advanced

Create the Filters

A filter is the actual file in which the values are defined. This may vary depending on which tool you’re using (Frame, Word, DITA, whatever), but the approach is generally the same between tools. As I mentioned in my last post, filtering is largely a mind-shift and less about which tool you’re using to accomplish the goal.

In my shop, we use DITA, so I would create a ditaval file for each of the cookbooks. (The filter files are called “ditavals” because they have a .ditaval filename extension).

  • Beginner
  • Advanced


Note that each of the ditavals includes/excludes the different audience values, like so:

  • Beginner — includes beginner, but excludes advanced.
  • Advanced — includes advanced, but excludes beginner.

The include/exclude action tells the computer which content to include and which to exclude when you publish your guide. So, for example, when you create the Beginner Guide, you want to see the following chapters:

  • Introduction
  • Cookies
  • Cakes and Pies
  • Basic Breads
  • Conclusion

You do not want to see:

  • Advanced Breads
  • French Pastry
  • Croissants

And when you create the Advanced Guide, you want to see:

  • Introduction
  • Cakes and Pies
  • Basic Breads
  • Advanced Breads
  • French Pastry
  • Croissants
  • Conclusion

You do not want to see:

  • Cookies

Tag at the Topic or Chapter Level

So, to accomplish this, you would “tag” the chapters so the computer knows what to show and what not to show at build/publish time. Again, in my shop, we use DITA, so my parent Cookbook file (called a “ditamap”) would look like this:


Breaking that Down

Introduction, Cakes and Pies, Basic Breads, and Conclusion are not tagged. That is because you want these to show (include) in both guides.

Here’s how the guides will look based on the include/exclude information in their associated filter file (the .ditaval):

  • Beginning Guide
    Will include Introduction, Cookies, Cakes and Pies, Basic Breads, Conclusion
    Will exclude Advanced Breads, French Pastry, Croissants
  • Advanced Guide
    Will include Introduction, Cakes and Pies, Basic Breads, Advanced Breads, French Pastry, Croissants, Conclusion
    Will exclude Cookies

If you don’t fully grasp this right now, try not to freak out. You’ll test it! You WILL get this to work. It just takes some time for your mind to adapt to this way of thinking.

Tag at the Inline Level

In addition to the topic-level tagging that you’ve done, you may have some text that you need to tag inline (e.g., specifying “audience=(value)” to specific elements). For example, in your Introduction chapter, you could do something like the following (apologies for crappy image; click it for better readability):


For the xml-phobic, here’s a different view of that same content with some nice red arrows to explain what’s going on (again, click the image for better readability):


Refer back to the “Breaking it Down” section above to understand how this tagged content will show or not show in your different cookbook versions.

QA the Tagging

Here are a few ideas for QAing your tagging:

  • Your tool may have a handy way to validate your tagging. In other words, if you have tagged something incorrectly, your tool may have a way to let you know that and show you exactly where you messed up. In my shop, we use Oxygen, which has a great validator.
  • You can also design in your own quick QA checks. For example, one thing I’ve done in the past is to create a topic that lists the PDF guides. Here’s an example where the table rows are tagged for either “sys_admin” or “end_user” audience value. In the final publication, I can quickly glance at this topic and see if the correct row is showing (in this case, my ditaval filters are such that only the end user row or only the system administrator row should be showing). Here’s how the XML looks:
    Documentation PDFs file in XML

And again for the xml-phobic, here is the same file in the WYSIWYG-ish view. The green indicates content that is tagged:
Documentation PDFs file in WYSIWYG

  • Compare the TOC of the final output against the filter. Is anything showing that shouldn’t be, or vice versa?
  • Look at your build output for any error messages. We’ll explore filtering challenges in another post, but for now, here’s a quick example of a filtering-related error message:
    [FATAL] Failed to parse the input file ‘AdministeringIdeation.ditamap’ because all of its content has been filtered out. Please check the input file ‘AdministeringIdeation.ditamap’ and the ditaval file, and ensure that the input is valid.

Create the Cookbooks

Now that you have tagged everything and you’re ready to create the guides, you’ll push the magic button in whatever tool you’re using (for example, running the build on the command line; running the build in Oxygen; running the build via your CMS, or whatever). In Oxygen, I would choose my filter (e.g., beginning_baking.ditaval), my desired outputs (HTML, PDF, .doc, whatever), and then run the build. I would expect the result to be the Beginning Guide.

Start Small! But Start!

I tried to pick an example that was complicated enough to describe the power of filtering, but still accessible enough that it wasn’t completely overwhelming. My advice would be to pick a small project and practice on it until you get the hang of filtering.

If you have any questions, feel free to ask in the comments. Thanks for reading!

July 23rd Meeting: Keyrefs with Roger Hadley

At our upcoming meeting on July 23rd, Roger Hadley of Grass Valley Software will explain keyrefs in DITA 1.2. You’ll learn what they are, how to use them (with real-world examples), and a few publishing pitfalls to avoid. All DITA experience levels welcome!

As usual, we’ll also do a little meet-and-greet, and enjoy a light meal together. We hope to see you.

Details: Wednesday, July 23, 2014, 6:30 – 8:00 p.m.

Jive Software, 915 SW Stark, Portland

Some Resources for Getting Started with DITA

I’ll be giving a five-minute lightning talk about DITA tomorrow at the Write the Docs meetup at Puppet Labs. Here are some resources for anyone considering using DITA:

Please post your own resources in the comments here so we can all benefit. Looking forward to the meetup tomorrow!