Not really. Remember, we're not talking about every step of the vetting, editing, and publishing process being automated; there's no way to do that, at least not using current technology, and so if your work passes the first hurdles it will still be considered by a human professional. It's only the first step of the process that could possibly be automated.
And so I can't really see the value in it. It's surely much cheaper to let an intern weed out the absolute rubbish, and forms a useful part of the education of a publisher.
The errors weeded out would be those that show a basic inability to write
- that's such a difficult concept to spec out. How do we define 'ability to write'?
and/or inability to follow correct query procedures: incorrect formatting
- we'd have to set it to only weed out egregiously bad formatting, because I don't want to filter out good things that have come out slightly off. The bar would have to be low.
- Impossible for a computer to assess in the foreseeable future - I can think of a lot of books it'd have a problem with even if you could work out a way of doing it. How can you algorithmically determine whether a book is SF or horror? Furthermore, how do you end up with new genres or with books that break genre rules?
certain tell-tale grammatical errors (the algorithm would probably need to make an exception when those errors occur inside quotes),
- Again, I have no idea what grammatical errors you could call 'tell-tale', and the ability to spot them implies, to my mind, some ability to parse the structure of natural language. And there are, again, plenty of books that make grammatical errors part of their style. (Trainspotting springs to mind. Do we have to write a Scots argot module for the filter?)
It's hard enough to write a regex that can 100% accurately determine which parts of a manuscript are inside quotes.
- excessive misspellings (ditto),
- Would have to set the bar pretty low again, I think, and it'd be hell on neologisms.
- If you can give instructions to a human being to do this, a programmer can give those instructions to a computer.
Well, I think there are a couple of assumptions there that just aren't borne out by the facts. The first is that the first filter is about stuff that an algorithm could possibly pick up. I don't read the first page of a manuscript looking for misspellings or grammar errors or the wrong genre. The second is that most of the things we can instruct a human to do cannot currently be emulated by software.
There's no guarantee of course that what passes the program's muster will be publishable, and no guarantee either that something good won't be weeded out by mistake on rare occasions, but that happens now, too.
True, but at least it happens according to human responses. We're trying to produce stuff that people will like, not robots.
I've seen programs that will write articles. Seriously. If that can be automated, so can the first steps of vetting a story. Also, so can the most basic editing.
I think the programs that write articles aren't working off an understanding of content, they're data mining tools. What you get is a summary of other articles