Sunday, October 12, 2025

Blue

Another plot bunny visited. This one is blue.

 

(From fanpop.com)

The first novel I finished, I had been writing an article for a gaming magazine. This was a little more straight-forward; I was thinking about concepts as a way to grasp cinematographic challenges of AI.

In that mix is thoughts I've had about wanting to do one of those Hornblower-esque career space navy things, and about wanting to do an engineer (who is more of a hacker), and thoughts about how to help Penny survive those situations where being Mistress of Waif-Fu would really, really come in handy.

And I ended up with a nice demonstration of how one idea can snowball into elaborate world-building...if you follow the potential implications. Start with a "heuristic implant." We're not going to get into the specifics of the tech here. Practically speaking, the young would-be soldier is sat under a helmet for a few hours, and when they get up;


Except experimental, very think-tank, with all the McNamara that promises. It is a whole set of combat skills that are deep-level muscle memory and happen basically automatically under the right stimuli. That right there is a whole host of problems. Bad enough your hands need to be licensed as deadly weapons -- now they are self-driving.

So that's a great character flaw, this killer instinct that could fire at the wrong moment, but at the same time, something that could help them through a sticky situation. Obviously (obviously, that is, when looked at from the needs of story, not through empty speculation on fantasy technology), they start meditating to at least control when it happens. And their relationship with this phantom driver...evolves.

(Yeah...Diadem from the Stars, The Stars My Destination, The Last Airbender -- not like this is exactly new ground.)

And that suggests a evolving situation, where a new kind of low-intensity warfare is challenging a military that organized entirely around capital ship combat. Of course that's a tired truism; the "always fighting the last war."

But that leads to wondering if this is really a navy at all. Or more like United Fruit Company, a massive corporate mercantile thing that works by rote and training and regulation, has been largely getting by with having a huge industrial base and leading-edge tech, but has expanded into a sector of space where the rules are a little different.

And now we've got multiple parties in the mix; old-school frontier traders who have the wisdom of experience, a cadre of experienced officers who want to create an actual military with a good esprit d'corps (not the same thing as free donut day at the office and "work smarter" posters in the cubicles), the friction between what is becoming an actual navy versus what is more like a merchant marine, the fresh-from-university theorists who pay far too much attention to how management thinks the world works (or should) and want quick-fix technological solutions over expensive training...

...and Ensign Blue in the middle of it, a trial run of one of the crazier outliers of the "super-soldier" package that various hard-liners have convinced themselves is the best way to win low-intensity conflict in an extremely politicized environment, a ship's engineer for a merchant ship who has no business at all getting hauled off to do dangerous missions on contested planets.

Oh, yeah. And engineer? Blue is the kind of engineer I've been seeing a lot of recently. Can (and often will) science the hell out of something, making the most amazing calculations. Then can't resist trying out a idea with ham-handed duct-tape and rat's-nest wiring that too often breaks (and sometimes catches fire).

Where does this go in character and career progression? Is there some third party, some out-of-context threat lurking behind what are still seen as basically raiding parties on the company's mining outposts? Are the rules about to go through another paradigm shift, tipping everyone from new enemies to unexpected allies into a brutal war?

Yeah, I got other books to write.

Tuesday, October 7, 2025

Anaconda

My go-to ComfyUI workflow now has more spaghetti than my most recent factory.


 (Not mine; some guy in Reddit.)

The VRAM crunch for long videos seems to rest primarily in the KSampler. There's an s2v workflow in the templates of a standard ComfyUI install that uses a tricky little module that picks up the latent and renders another chunk of video, for all of them to be stitched together at the end. With that thing, the major VRAM crunch is size of the image.

Of course there's still the decoherence issue. I've been running 40-second tests to see how badly the image decomposes over that many frames. Also found the quality is acceptable rendering at 720 and upscaling to 1024 via a simple frame-by-frame lanczos upscaler (nothing AI about it). And I'm rather proud I figured that our all by myself. At 16 fps and with Steps set down at 4 I can get a second of video for every minute the floor heater is running.

Scripting is still a big unknown. I've been experimenting with the s2v (sound to video) and as usual there are surprises. AI, after all, is an exercise in probabilities. "These things are often found with those things." It is, below the layers of agents and control nets and weighting, a next-word autocomplete.

That means it seems to have an uncanny ability to extract emotional and semantic meaning from speech. It is strictly associational; videos in the training material tended to show a person pointing when the vocal patterns of "look over there" occurred. More emergence. Cat logic, even.



So anyhow, I broke Automatic1111. Sure, it had a venv folder, but somehow Path got pointed in the wrong direction. Fortunately was able to delete python, clean install 3.10.9 inside the SD folder, Automatic1111 came back up and ComfyUI was still safe in its own sandbox. And now to try to install Kohya.


Experimenting with the tech has led to thinking about shots, and that in turn has circled back to the same thing I identified earlier, a thing that becomes particularly visible when talking about AI.

We all have an urge to create. And we all have our desires and internal landscapes that, when given the chance, will attempt to shape the work. Well, okay, writing forums talk about the person who wants to have written a book; the book itself being of no import, just as the nature of the film they starred in having nothing to do with the desire to be a famous actor. It is the fame and fortune that is the object.

In any case, the difference between the stereotype of push-button art (paint by numeric control) and the application of actual skills that took time and effort to learn is, in relation to the process of creation itself, just a matter of how granular you are getting about it.

Music has long had chance music and aleatoric music. Some artists throw paint at a canvass. And some people hire or collaborate. Is a composer not a composer if they hire an arranger?

That said, I feel that in video, the approach taken by many in AI is getting in the way of achieving a meaningful goal. As it exists right now, AI video is poorly scriptable, and its cinematography -- the choice of shots and cutting in order to tell the story -- is lacking. This, as with all things AI, will change.

But right now a lot of people getting into AI are crowding the subreddits asking how to generate longer videos.

I'm sorry, but wrong approach. In today's cinematography, 15 seconds is considered a long shot. Many movies are cut at a faster tempo than that. Now, there is the issue of coverage...but I'll get there. In any case, this is just another side of the AI approach that wants nothing more than to press buttons. In fact, it isn't even the time, effort, or artistic skills or tools that are being avoided. It is the burden of creativity. People are using AI to create the prompts to create AI images. And not just sometimes; there are workflows designed to automate this terribly challenging chore of getting ChatGPT to spit out a string of words that can be plugged into ComfyUI.

Art and purposes change. New forms arise. A sonnet is not a haiku. There is argument to recognize as a form the short-form AI video that stitches together semi-related clips in a montage style.

But even here, the AI is going to do poorly at generating it all in one go. It will do better if each shot is rendered separately, and something (a human editor, even!) splices the shots together. And, especially if the target is TikTok or the equivalent, the individual shots are rarely going to be more than five seconds in length.


Cutting to develop a story, using language similar to modern filmic language, is a different beast entirely. The challenge I'm thinking a lot about now is consistency. Consistency of character, consistency of set. There are also challenges in matching camera motions and angles if you want to apply the language correctly. For that shot-reverse-shot of which the OTS is often part, you have to obey the 180 rule or the results become confusing.

One basic approach is image to video. With i2v, every shot has the same starting point, although they diverge from there. As a specific example, imagine a render of a car driving off. In one render, the removal of the car reveals a fire hydrant. In the second render from the same start point, a mailbox. The AI rolled the dice each time because that part of the background wasn't in the original reference.

One weird problem as well. In editing, various kinds of buffer shots are inserted to hide the cuts made to the master shot. The interview subject coughed. If you just cut, there'd be a stutter in the film. So cut to the interviewer nodding as if listening (those are usually filmed at a different time, and without the subject at all!) Then cut back.

In the case of an i2v workflow, a cutaway done like this would create a strange déjà vu; after the cut, the main shot seems to have reset in time.

So this might actually be an argument for a longer clip, but not to be used as the final output; to be used as a master shot to be cut into for story beats.

Only we run into another problem. It is poorly scriptable at present. In the workflows I am currently using, there's essentially one idea per clip. So a simple idea such as "he sees the gun and starts talking rapidly" doesn't work with this process.

What you need is to create two clips with different prompts. And you need to steal the last frame from the first clip and use it as the starting image of the second clip. Only this too has problems; the degradation over a length of a clip means even if you add a node in the workflow to automatically save the target frame, it will need to be cleaned up, corrected back to being on-model, and have the resolution increased back to the original.

And, yes, I've seen a workflow that automates all of that, right down to a preset noise setting in the AI model that regenerates a fresh and higher-resolution image.

My, what a tangled web we weave.

Monday, October 6, 2025

Cryptic Triptych

I got the PC I built up and running, after the usual 22h2 hassle (tip; don't use the internal updater. Run the web installer at Microsoft. For as long as that lasts!)

ComfyUI is sandboxed (and a one-click install) and Automatic1111, though now an abandoned project, also installs a venv folder within the stable_diffusion folder, meaning it can run on Python 3.10.6  Now trying to get Kohya running, and learning venv so I can get that on 3.10.11 or higher...without breaking everything else.

I still like the primitive but functional GUI of Automatic1111 for stills. But ComfyUI opens up video as well. Motion.

And that got me thinking about linear narrative.

There does exist a form call "non-linear narrative." But that refers to the relationship between the narrative and some other chronology. The latter may be shifted around. A writer can at any point refer to a different time, including such techniques as the flashback and flash-forward. But the narrative itself remains linear. One reads one word at a time.

(Arguably, from our understanding of the process of reading, we parse chunks of text and thus multiple words may be included in what is experienced as a phoneme of extracted meaning.)

This means it is extremely difficult to capture the near-simultaneous flow of information that a real person not reading an account in a book would experience. In our old gaming circles, the joke was the monster quietly waiting until the room description was finished. It is a basic problem in writing; you can't tell it all at the same time. And the order you chose influences the relative weight given.

Again, arguably, our attention can't be split too many ways. In most cases, the realization you had while you were in the middle of doing something else arrives as a discrete event. You may have heard the voice behind you, and been processing it, but the moment of understanding that cry of "Stop!" can be treated narratively at the moment it becomes the focus of attention. And the observations that lead to that moment of realization back-filled at that time, as they, too, rise to the top of the consciousness.

Or in another way of putting it, a narrative is an alias of the stream of consciousness and the order of presentation can be taken as the order of items brought into focus.

This idea of the sequential scroll of attention has been used in artwork. We normally absorb a piece of art by moving from one focus to another (in a matrix of probable interest including size, color, position, human faces, etc.) The artist can construct a narrative through this shifting of focus.


This one sneaks up in stages. The first impression is very calming. The next impressions are not. Especially in some periods, there could be subtler and subtler clues and symbols that you don't notice until you've been looking for a while.


Or there are artists, from the triptych to the Bayeux Tapestry that arrange distinct framed panels in a sequential order.

Motion controls this flow of narrative more tightly. Not to say there can't be the same slow realizations. But it means thinking sequentially.


In comic book terminology, the words "Closure" and "Encapsulation" are used to describe the concepts I've been talking about. "Closure" is the mental act of bringing together information that had been presented over a sequence of panels in order to extract the idea of a single thing or event. "Encapsulation" is a single panel that is both the highlight of and a reference pointer or stand-in for that event.

In text, narrative, especially immersive narrative that is keyed to a strong POV or, worse, a first-person POV, has a bias towards moving chronologically. Especially in first-person, this will lead the unwary writer into documenting every moment from waking to sleep (which is why I call it "Day-Planner Syndrome.")


I've been more and more conscious of the advantages and drawbacks of jumping into a scene at a more interesting point and rapidly back-filling (tell not show) the context of what that moment came out of. I don't like these little loops and how they disturb the illusion of a continuous consciousness that the reader is merely eves-dropping on as they go about their day, but I like even less spending pages on every breakfast.

And speaking of time. The best way to experience the passage of time is to have time pass. That is, if you want the reader to feel that long drive through the desert, you have to make them spend some time reading it. There's really no shortcut.

I decided for The Early Fox I wanted to present Penny as more of a blank slate, and to keep the focus within New Mexico. So no talking about her past experiences, comparisons to other places she's been, comparison or discussion of the histories of other places, technical discussions that bring in questions of where Penny learned geology or Latin or whatever, or quite so many pop-culture references.

And that means I am seriously running out of ways I can describe yucca.


In any case, spent a chunk of the weekend doing test runs with WAN2.2 and 2.1 on the subject of "will it move?" Which is basically the process of interrogating an AI model to see what it understands and what form the answer will take.

My first test on any new install is the prompt "bird." Just the one word. Across a number of checkpoints the result is a bird on the ground, usually grass. A strange and yet almost specific and describable bird; it is sort of a combination of bluebird and puffin with a large hooked beak, black/white mask, blue plumage and yellow chicken legs.

In investigating motion in video, I discovered there are two major things going on under the hood. The first is that when you get out of the mainstream ("person talking") and into a more specific motion ("person climbing a cliff") you run into the paucity of training data problem. When there is a variety of data, the AI can synthesize something that appears original. When the selection is too small, the AI recaps that bit of data in a way that becomes recognizable. Oh, that climbing move where he steps up with his left foot, then nods his head twice.

The other is subject-background detection. AI video works now (more-or-less) because of subject consistency. The person walking remains in the same clothing from the first frame to the last. It does interpolate, creating its own synthesized 3d version, but it can be thought of as, basically, detaching the subject then sliding it around on the background.

We've re-invented Flash.


Now, because the AI is detaching then interpolating, and the interpolation makes use of the training data of what the back of a coat or the rest of a shoe looks like (and, for video models, moves like), it does have the ability to animate things like hair appropriately when that subject is in motion. But AI is pretty good at not recognizing stuff, too. In this case, it takes the details it doesn't quite understand and basically turns them into a game skin.

Whether this is something the programmers were thinking, or an emergent behavior in which AI is discovering similar ways of approximating reality to what game creators have been doing, the subject becomes basically a surface mesh that gets the large-scale movements right but can reveal that things like the pauldrons on a suit of armor are basically surface details, parts of the "mesh."

It can help to think of AI animation as Flash in 3D. The identified subjects move around a background, with both given consistency from frame to frame. And think of the subject, whether it is a cat or a planet, as a single object that can be folded and stretched with the surface details more-or-less following.

But back to that consistency thing. For various reasons, video renders are limited to the low hundreds of frames (the default starter, depending on model, is 33 to 77 frames). And each render is a fresh roll of the dice. 

It is a strange paradox, possibly unavoidable in the way we are currently doing this thing we call "AI." In order to have something with the appearance of novelty, it has to fold in the larger bulk of training data. In order to have consistency, it has to ignore most of that data. And since we've decided to interrogate the black box of the engine with a text prompt, we are basically left with "make me a bird" and the engine spitting out a fresh interpretation every time.

That plays hell on making an actual narrative. Replace comic-book panel with film-terminology "shot," and have that "Closure" built on things developed over multiple shots, and you are confronted with the problem that the actors and setting are based on concepts, not on a stable model that exists outside the world of an individual render. If you construct "Bird walking," "Bird flies off," and "Bird in the sky" with each time interpreting the conceptual idea of "Bird," in a different way it is going to be a harder story to understand.


That is going to change. There are going to be character turn-arounds or virtual set building soon enough. As I understand it, though, the necessary randomness means the paradox is baked into the process. No matter what the model or template, it is treated the same as a prompt or a LoRA or any other weighting; as a suggestion. One that gets interpreted in the light of what that roll of the dice spat out that run.

And that's why the majority of those AI videos currently clogging YouTube go for conceptual snippets arranged in a narrative order, not a tight sequence of shots in close chronological time. You can easily prompt the AI to render the hero walking into a spaceport, and the hero piloting his spacecraft...now wearing a spacesuit and with a visibly different haircut.

For now, the best work-around appears to be using the "I2V" subset. That generates a video from an image reference. The downside is that anything that isn't in the image -- the back of the head, say -- is interpolated, and thus will be different in every render. It also requires creating starter images that are themselves on-model.

A related trick is pulling the last frame of the first render and using that as the starter image for a second render. The problem this runs into is the Xerox Effect; the same problem that is part of why there is a soft limit of the number of frames of animation can be rendered in a single run.


(The bigger problem in render length is memory management. I am not entirely clear why, 

As with most things AI, or 3D for that matter, it turns into the Compile Dance. Since each run is a roll of the dice, you often can't tell if there is a basic error of setup (bad prompt, a mistake in the reference image, a node connected backwards) or just a bad draw from the deck. You have to render a couple of times. Tweak a setting. Render a couple times to see if that change was in the right direction. Lather, rinse.

With my new GPU and the convenient test size I have been working with, render times fall into the sour spot. 1-3 minutes; not long enough to do something else, but long enough it is annoying to wait it out.

I still don't have an application, but it is an amusing enough technical problem to keep chasing for a bit longer. The discussions on the main subreddit seem to show a majority of questioners who just want "longer video" and hope that by crafting the right prompt, they can build a narrative in an interesting way.

The small minority is there, however, explaining that cutting together shorter clips better approaches how the movies have been doing it for a long time; a narrative approach that seems to work for the viewer. But that really throws things back towards the problem of consistency between clips.

And that's why I'm neck-deep in Python, trying not to break the rest of the tool kit in adding a LoRA trainer to the mix.

Thursday, October 2, 2025

How does your garden grow

Sometimes scenes evolve. You realize things the outline didn't address, and addressing them opens up opportunities you hadn't spotted before.

And, I know. The fierce outliners and the dedicated pantsers (and especially the WriMoNo -- NaNoWriMo is dead but some people still try to do 50K over November) call you stupid if you delay in getting those words on the page. But I think that time comes due eventually. Fix it now, or fix it in rewrites. As long as you are capable of holding enough story in your head, you don't have to have it written down in order to realize it is wrong and needs to be re-written.

The drive from Alamogordo to the tiny census-designated place (it's too small to be a town) of Yah-ta-hey is about five hours. The original plan/outline/mind mapper diagram just said "Mary leads Penny to the rez to talk to a man who knows about the 'Sheep Ranch'."

The original spec was that Penny doesn't get friends in this one. She doesn't get people she can lean on for emotional support, or people who are too helpful. They all have agenda, problems, and hide stuff from her. Mary is mostly...angry. I got her character and voice from a couple of different writers on continuing problems with radiological contamination on tribal lands.

But can she and Penny take a five-hour drive without either getting some resolution, or killing each other? Am I better off having Penny go alone? Or...is it the better option to make this a longer scene, to go deeper into Mary's personality, and give them a mini character arc?

Sigh. The hard one, of course.


The world continues to change. I started this in 2018. An American tourist loose in the world. The world's view of America has changed since then. COVID has changed things, the economic slump has changed things, and tourism has changed for everyone; European tourists are in the same hot water with over-travelled destinations and Venice is not the only place fighting back.

It always blind-sides you, change does. I had minor bit on a minor characters; a senior airman at the 49th with a shaving profile. I couldn't even name it, because that's the sort of thing that people who have been in the service would recognize but most of my readers would never have heard of.

Until it suddenly became a big thing to the pushup king, our current Secretary of War.


I did happen to think of a new silly idea, possibly for the "weird high-tech company in the wilds of Colorado" next adventure. Probably more productive is I've finally started making a proper vector of the new series logo so I can try out the new cover.

So there actually is a magical artifact. Sort of not very. It is extremely valuable and holds secrets because it is a replica and there's a USB stick in it with a bunch of trade secrets or something.

And maybe, I thought, it isn't the product of a historical culture, but instead a replica prop. And comes out of an imaginary IP, some sprawling fantasy saga with a lot of borrowing from various bits of real history (like GOT, say).

I mentioned a while back that various authors have tried to create a Disneyworld based on an imagined IP. The tough part is communicating this IP to the reader; you can't just say, "Look, it's Elsa from Frozen!" The fun part is, of course, creating the IP in the first place.


And, yeah, finally broke down and built a new computer. Things have moved on there, as well. SSDs aren't being fitted with expansion rails so they can slot in where the old 3.5" drive bays are. Instead they have a PCIe slot on the motherboard -- underneath a built-in heatsink because modern gaming machines run hot.

Darkrock case, MSI board, I7 CPU, 64gb starter RAM, 2tb SSD as the "C" drive, 1000 watt PSU and of course a nice floor heater of an RTX 3090 with 24gb of VRAM. 

Mostly did it because I wanted to do it. Not because I had a need for it. But, oh boy, when I finally got the updates installed and running and tried out Satisfactory it looked so good...

The test render on WAN2.2 t2v took 160 seconds. Not bad.

Tuesday, September 23, 2025

Remember the Espirito Santo

I kinda want to send Penny to Texas next. They got oil, some very old towns, the Alamo, and about a million interesting wrecks off the coast. They also got more dinosaurs than you can shake an animatronic thagomizer at.


 Goings-on at a fancy dinosaur exhibit is fun. Everybody loves some amusement park shenanigans, and adding dinosaurs is only further insurance that something will go worng.

Of course archaeologists famously don't do dinosaurs.


But I'm not all that enthralled by doing archaeology on missions, old towns, or the Alamos, or even Washington-on-Brazos, the "birthplace of Texas."


I'm more intrigued by marine archaeology. Always wanted to work that in at some point. But that's a ton of research. And, really, the good way to phrase "Write what you know" is "Write what you want to know." Write about things that excite you, either through prior exposure or because you really want to write about them.

I've got this emotional stub of Penny being a low-wage worker at some soul-less tech center in some up-and-coming development on the outskirts of a national park...until something goes terribly right. (Wrong for her, right for getting a story out of it.)

Well, I really do have a prior novel to finish. Then I might clean the decks, stick a plastic fern in the corner and hang some paper lanterns, and get the hell out of First Person POV for a while.

Assembled a cart at NewEgg, too. I can only sorta write during downtime at work and looking up parts -- as annoying as it can be -- is a way to pass the time. Of course, the only graphics project I've got that is actually for anyone other than myself, is the new cover I'm trying...and there's no AI involved in that.


Thursday, September 18, 2025

Painting with a potato

There is an artist's method. Like scientific method or engineering method, it is a base process. This idea of processes that are largely application-independent underlies the Maker movement -- which has merged and morphed and now is more like the kind of "hacker" the Hackaday Blog was built for.


This is why there are so many multi-instrumentalists in the home studio music circles, and how visual artists change media. There's cross-over, too. Not to in any way diminish the immense investment in skills in the specific mediums and tools, but there are commonalities at more fractal levels than just the generalized artistic method.

In any case, the thinking is basically tools is tools. I've designed sound and lighting for the stage. I've designed props. I've done book covers. I've done music. I'm not saying I do any of these things well, but there is a hacker approach to it. Understand the new medium and tools and then apply the core skills to it. It all feels sort of the same, reaching for personal aesthetics and understanding of the real world and learning about the tools and the traditions of this specific media.

From a process point of view, from this higher level "making art" point of view, AI is just another tool set.

It tends to be a toolset that is relatively easy to learn, but also less responsive. It really is like painting with a potato (and not to make stamps, either!) You can't really repurpose the hand-and-eye skills learned in inking a comic book panel with a nibbed pen or painting a clean border on a set wall with a sash brush. Because it is a potato. You hold it, you dip it in paint, it makes a big messy mark on the canvas.

It is mostly the upper-level skills and that general artistic approach that brings to bear. What of the things I can imagine and want to visualize, will this tool let me achieve? You can't get a melody out of a sculpture or paint the fluting on a hand-knapped obsidian point, and you learn quickly what subjects and styles and so forth the AI tools can support, and which are not really what they are meant for.

This will change. This will change quickly. Already, I've discovered even my potato computer could have done those short videos with animation synchronized to the music.

That's been the latest "Hey, I found this new kind of ink marker at the art store and I just had to try it out on something" for me. S2V with an audio-to-video model built around the WAN2.2 (14B) AI image model. In a spaghetti workflow hosted in the ComfyUI.

On a potato. (I5 with 3060Ti. I want to upgrade to 3090 -- generations don't matter as much as that raw 24GB of VRAM -- but can't stomach another thousand bucks on the proper host of an I9 machine with the bus for DDR5.)

Render times for ten seconds of video at HD (480x640) is around an hour. Hello, 1994; I'm back to Bryce3D!


Actually, right now longer clips lose coherence. That's the reason all that stuff on YouTube right now is based on 30 seconds and under. It's the same problem that underlies why GPT-4 loses the plot a chapter in to a novel and ACE-step forgets the tune or even the key. 

I suspect there's a fix with adding another agent in there so it won't surprise me to find this is a temporary problem. But it also won't surprise me if it proves more difficult than that, as this is out of something that sort of underlines the entire genetic algorithm approach of things. Even on a single still image, I've seen Stable Diffusion forget what it was rendering and start detailing up something other than what it started with. Particularly a problem with inpainting. "Add a hand here." "Now detail that hand." Responds SD "What hand?"

(This also may change with the move towards natural language prompting. The traditional approach was spelling things out, as if the AI was an idiot. This is the "Bear, brown, with fur, holding a spatula in one paw..." approach. The WAN (and I assume Flux) tools claim better results with "A bear tending a barbecue" and let the AI figure out what that entails.)

Anyhow. Been experimenting with a wacky pipeline, but this time using speech instead of music.

Rip the audio track off a video with VLC. Extract and clean up a voice sample via Audacity. Drop voice sample and script into Chatterbox (I've been messing with the TTL and VC branches.) The other pipeline is using basic paint to compose, Stable Diffusion (using Automatic1111 as a front end, because while it may be aging, the all-in-one canvas approach is much more suitable for inpainting. Plus, I know this one pretty well now.)

Throw both into the WAN2.2-base S2V running under ComfyUI. It does decent lip synch. I haven't tried it out yet for instrumentals but apparently it can parse some of that, too. It also has a different architecture than the WAN2.1 I was doing I2V on before. It has a similar model consistency, which is a nice surprise, but the workflow I'm using leverages an "extend video" hack that means that my potato -- as much as it struggles fitting a 20GB model into 12GB of VRAM -- can get out to at least 14 seconds at HD.

As usual, the fun part is trying out edge cases. Always, the artist wants to mess around and see what the tools can do. And at some point you are smacking that sculpture with a drum stick, recording the result and programming it into a sample keyboard just to see if it can be done.

There is an image-to-mesh template already on my ComfyUI install. The AI can already interpolate outside of the picture frame, which is what it is doing -- almost seamlessly! -- in the image2video workflow. So it makes sense that it can interpolate the back side of an object in an image. And then produce a surface and export it as an STL file you can send to your physical printer!

So there are many things the tools can do. The underlying paradox remains -- perhaps even more of one as the tools rely on the AI reading your intent and executing that; an understanding that relies on there being a substantial number of people that asked for the same thing using similar words.

It is the new digital version of the tragedy of the commons. It is the search algorithm turned into image generation. It has never been easier to look like everyone else (well, maybe in Paris in 1900, if you were trying to get selected for show by the Académie).

And that indeed is my new fear. I think there is this thought in some of the big companies that this could be a new way to stream content. Give everyone tools to "create art" so, like the nightmare the Incredibles faced, nobody is special. The actual artists starve out even faster than they have been already. The consumer believes they are now the creator, but they aren't; they are only the users of a service that spits out variations on what they have already been sold.

Right now, AI art is in the hacker domain, where academics and tinkerers and volunteers are creating the tools. Soon enough, the walls will close. Copyright and DMCA, aggressive "anti-pornography" stances. None of that actually wrong by itself, but applied surgically to make sure the tools are no longer in your hands but are only rented from the company that owns them.

The big software companies have often been oddly hostile to content creators. "Why would you need to create and share when we've made it so easy to watch DVD's?" This just accelerates in that direction, where the end-user doesn't create and doesn't own.

We're back to streaming content. It is just the buttons have changed labels. Instead of "play me something that's like what I usually listen to" it will be, "create me something that sounds like what I usually listen to.

Doesn't stop me from playing around with the stuff myself.


(Call to Power players will recognize this one.)

Monday, September 15, 2025

It's all about ethics in archaeology

I've a writing worklist right now.

Finishing The Early Fox is top of the list, of course. I didn't find that Civil War graveyard (there are a lot of "Billy the Kid kilt heah" locations but nothing near enough to US-54) but I did find an old Mission style (WPA project, actually) school in a gold boom ghost town. That will do for my "Ecstasy of Gold" scene.

Still not ready for the foray into Navajo lands. 

Contacted the cover artists I've worked with before but they were uninterested in doing a cover clean-up. They have (most of these guys have) a listed service of cover doctor, but what they mean is, they'll design one for you to replace whatever it is you are using.

So I guess I'm doing my own cover art again. After my last couple of go-arounds, I'm tired of struggling to communicate. And first task on that list is doing my own vectors over the art my logo designer did. And probably changing a few things and trying out a few things, too.

As part of the rebranding, revisions to The Fox Knows Many Things. Including AI -- but not what you'd think! ProWritingAid has an AI development editor thing that is, of course, extra fees but they have a free trial. That's not in any way using AI to revise.

It is, perhaps, one of the best ways of using AI; a thousand-monkey version of crowdsourcing (on stolen data). Point the AI monkey at the manuscript and see if enough of it fell within the kinds of things readers are used to seeing (aka looks like other books) that the AI can figure out what it is looking at.

Always, as with any editing tool or feedback, take it as a suggestion and apply your own instincts and intelligence to those suggestions. But I'd be interested in seeing what it comes up with.

***

I've been thinking about archaeological issues to hang the next book on.

That is my conceit for the series; each book features a place and culture, a bit of history connected to said place and culture, and explores some aspect of modern archaeology.

The current book, I backed off early with the idea of doing "bad experiences on digs." The sexism on the dig in this book is a plot point, but it isn't defining the dig or Penny's experience (which is largely positive...next book, maybe, I can get her the Job from Hell.)

So The Early Fox is more-or-less NAGPRA and associated issues. I was barely able to touch on indigenous archaeology/community archaeology, the mound builder myth, and so on but at that, it did better than the Paris book in having something to do with archaeology.


Over previous books, I've already delved into archaeological looting, nationalist revisionism and irredentism. And pseudo-archaeology, but there are too many easy pickings there. I could easily fill a series with just running around after the Kensington Runestone or Bosnian pyramids. The fake artifact trade is interesting but I've got a book saved for that. And repatriation issues show up again with the boat episode (assuming I get around to writing either of those).


And somehow that thinking led to something that was an actual excuse other than "this is the random small town I ended up in next" for Penny's involvement. Brought in to help the artifacts look more authentic (for extremely suspect values of "authentic") in some kind of weird high-tech VR sort of thing. Less gaming company (why would they need a fancy tech center?) more experimental direct-mind-connection stuff.

Not really directions I wanted to take the series, but then having gunfights with ISIL over some Buddhist statuary isn't where I want to go, either.

Thursday, September 11, 2025

Experimenting with ACE-step


Yeah, I've gone to the dark side.

I said I was messing around with AI music creation via ACE-step. That one is a bit more towards a diffusion model, so (at the moment) less controllable than Suno, but potentially more flexible. They boast about the number of languages it can handle...but at the same time, admit that they really haven't trained it on opera, or (as my own small experiments seemed to show) much of anything at all outside of pop music.

(They also side-step around the sound quality, with the user base developing magic incantations that might be improving things if you squint, or might just be random "sometimes it doesn't sound quite so bad" and has nothing to do with their "and set cfig to 1.14" stuff.)

The one that I actually wanted to talk about was the attempt to do a villain song/"I want" song. The result is less Broadway, more Billie Eilish. Among other things, ACE doesn't know how to belt. Or like female singers in a range other than light soprano. It's understanding of "chorus" v. "verse" is "sing louder!" It also only takes lyrics, like anything else, as suggestions. 

But on the fourth or fifth run it came out with this weird melisma that's almost Sondheim or something. Pity it couldn't carry through. I could improvise a better chorus (it is an absolute natural to sequence this one, half-stepping higher and higher). Plus, at least one previous run was smart enough to do the intro on solo piano. I like the "is that supposed to be an oboe?" but it should enter later.

Like I said. Not really trained on showtunes. That's why it surprised me that this run, instead of repeating the chorus over and over, drifting further and further from the melody as it goes (and not even a Truck Driver Gear Shift!) it went for an extended instrumental break. Cool. On stage, this would be a dance, or perhaps an aerial dance.

The less said about the Wan2.1 animations, the better. At some point in the increasingly large and power-hungry tools you can script better, and start using LORA to keep the animator from going completely off model. At this level it is just as well I've got a 3060 ti with 12Gb that is practically a potato when it comes to video -- because Wan2.1 also loses the plot after 60-90 frames of video. So no Bryce-length render times. It also has some memory leaks so even if I could automate a batch, the second one would fail.

But on the other hand; this gave a nice chance to demonstrate using img2img mode for composing a Stable Diffusion image. I did those roughs in Paint3D, using my knowledge of how SD "sees" an image. The airship took a bit of inpainting as the AI really wanted to make either a blimp or a clipper, not try to get both in the same machine. The others I used simple convergence; rendered a half-dozen versions, picked one, then re-rendered on that one to zero it in.

At least with SDXL you can prompt a little. ACE-step and WAN2.1 go even more nuts about stray words -- my first attempt at the airship had "flying" in the prompt and the AI added wings to it. For the ACE renders, my prompts were as short as four words long.

The other lesson here is that the AI will willingly change lots of things, but it is real stickler about proportions. What it saw in the comp, it will keep. A big-headed guy with a short-necked guitar is going to stay that way through the generations, and it is really difficult to fix it once you are in the clean-up render phase.


Wednesday, September 10, 2025

Fox and goose

Spent a chunk of the weekend wrestling with mostly that class of problem. Now I'm outlined to the end. I know; I was outlined before. Now I have more detail.


For me, it isn't figuring out why there's a bomb on the train. Or how the heroes figure out that there is a bomb on the train. It is all about controlling when they find out. That's the part that annoys me and feels like it takes a lot more time than it should.

That's because the why is rooted in the scene in which that clue is dropped, and the ordering of the scenes is at the mercy of geography, time, and all the other plot threads that led to them finding the briefcase full of blasting caps. Like, the briefcase was dropped by the Crooked Man during the chase scene, but that means he was heading away from the station, so how did he end up back on the train?

Multi-value problems. 

Weirdly, this wraps all the way back to one of the basics. Having a bat-computer moment, "I just realized there's a bomb on the train!" is telling. You want the moment that information arrives to be dramatic, and that is practically synonymous with showing it.

Which is to say, it is the central business of a scene. Which makes the scenes inseparable from the clues, which means you have to figure out how to do them in the one order that doesn't end up with a fat chicken and an even fatter fox.


My concentration wasn't up for that for most of the last few days so I made progress on the new cover/series rebrand. I'd roughed out a logo but I didn't quite have the patience for dragging vectors around. And the font I used ("Adventure") isn't licensed for commercial use.

Took a chance on a guy in Fiverr who did not quite share enough English with me to make communication fluid, but seemed familiar with the style I wanted. Nobody in this business makes sketches anymore. The worst are the cover designers. They give you a finished cover and besides the limited number of revisions they allow, you feel bad about asking them to cut into a finished artwork.

That is because of the pipeline, which is heavily asset-based, and I understand. There's nothing wrong about having a cover that looks like other covers in the same genre. It makes it easier for the reader to find the book, and it give the reader a certain trust that they are buying the kind of book they wanted to be buying.

It can lead to asset over-use, though.


My Fiverrr guy was showing me full-color graphics. So, fine, maybe he liked sketching with paint tools. The order was explicitly for vector graphics, though. Was he going to turn this more cartoony paint into proper vectors at the last step?

Turned out, that was the finished art. The vectors were him throwing blobby paint sketch into auto-convert, making one of those impossible-to-edit vector "bit clouds."

So I'm doing the vectors after all. Sigh. Well, half the reason I hired an outsider is because I'm at best a craftsman and I wanted the imagination of an artist. And I liked his ideas. But that's left me even less eager to hire other people.

Because you are up against efficiency. They don't want to do your ideas; they want to do the idea that can be done with stock assets, quick-and-dirty "convert to vector," and more and more, AI. And on the way to talking you into picking that instead (via making a really glossy final version they hope you will be attracted to enough to go with despite reservations), talk you up into Facebook banner and animated logo and an AI image of a hand holding your book.

I already dropped a hundred on a cover I don't want to use. I may just have to go back to doing my own.

Monday, September 8, 2025

In a cavern in a canyon...

...excavating for bitcoins?

I'm not really where I want to be as a writer to tackle the New Orleans story. But more to the point, I've got this weird attraction towards the idea of sending Bill Bixby///Penny to Colorado or Montana or something, find a low-population bit of near-wilderness and put a massive super-advanced tech center for her to get in trouble in.

So, like, that new Google AI center, only located in an old coal mine. Or something else suitably techno-thriller epically over-scaled and real-world commercially implausible.



I can find a remote mountain or a salt mine or something else sufficiently crazy. I can invent a biotech firm or a nanotech lab or some other appropriately dangerous and dystopian technology.

What I'm having trouble with, is tying it to the conceits of the Athena Fox series. Which are that there is a subject of history somewhere in the mix (preferable a narrow or easy-to-define band in time and/or space). That there is an archaeological issue to raise, like archaeological ethics or the problem with field schools. And that at some point it goes Tomb Raider, in that there is some perfectly sensible reason why Penny finds herself crawling down booby-trapped tunnels or fighting with ninja.

For the Bill Bixby stories, she doesn't need a good excuse to be there. There's no dig involved. She got off the Greyhound somewhere in middle America. Or got a phone call.


Yeah, like I'm in a good place to be plotting a new book. I'm having shiny new "I'd rather be writing anything but this" syndrome. Actually, I feel okay about The Early Fox. It went different than I expected, but it looks like I'm hitting the beats I hoped to hit.

I'm enjoying writing it, it is just a lot of work and I'm having trouble coming up with the sheer concentrated brainpower to deal with the stuff that's outlined.

The Atlas-F silo won't be bad, I can stumble through the next Freeman conversations, the long chase down the Jornada de Muerto has complexities but it isn't anything I haven't done before.

No, but going to the rez is challenging. Not looking forward to tackling that. Worse, I decided I need to send her to the War Zone (to contact an urban explorer who can tell her about Lon's errand at the Atlas-F silo outside of Roswell). Which almost certainly includes cops, and that means the deputy sheriff at Alamogordo is firmly pencilled in now. 

And I still have to go back and do the nuke museum scene. The museum is fine. The "Los Alamos Wife" bit is going to take a little more. Probably not watching a full season of Manhattan. Fortunately, the museum has free online videos that are somewhat more...historical.




Saturday, September 6, 2025

I, for one, welcome our new...

...clipping service?

Yes, AI is a fancy plagiarism machine. But that doesn't make it entirely useless as a sort of clipping service; it goes out there and finds things that are like what you seem to be describing, shamelessly rips them off and shows you the resulting pastiche.

Which you then get independently inspired by and do your own, only vaguely similar, thing in response.

Since I went ahead and put ComfyUI -- it of the many nodes -- on my gaming machine (it's the one with the fast graphics card), I tried out text-to-music.

I am not impressed. Besides all the usual artist-type comments about how it doesn't understand anything about rhythm or harmony, and the front end was written by someone who doesn't know a fugue from a flugelhorn (and may have never even heard of the latter), the technical side of me was appalled.

The sound quality sucks rocks. It is damned near unlistenable.

Still, a couple of runs with the prompt "steampunk sea chanty" were amusing to listen to and tempted me for a long moment to go break out the more traditional tools.

This particular implementation (not that ACE is getting top marks in anyone's text-to-music books) totally sucks rocks. And not in a good way.


And this is technically interesting from two totally different directions. One is that getting a glossy appearance is one of the things AI is really good at (too good; it gets called out for that too-perfect "AI look" all the time). So why is ACE so bad? And is this endemic across using diffusion models for music creation?

For that matter, are there competing technologies and if so, what are they? I mean, we've been doing computationally generated music since before there were computers. With the robust ecosystem of software synthesis and sound manipulation, attacking the problem Band in a Box style (which was a pattern-fill sort of MIDI generator) seems more fruitful.

The other is that this seems so very...useless. Drawing is hard. It takes a lot of time to learn and even if you are good and using tools that are fast (some of which highly leverage computing in various ways) it still takes time to do.

Punching a keyboard controller is so much less effort in comparison. The barrier to music creation at home, now that there are things like Garage Band being given away with the operating system, seems very short to me.

But then, isn't that what AI is basically making a name for itself with? Taking tasks you didn't really need to automate, and automating them anyhow?

***

In more productive news, I made a full mock-up of my "map with red line" cover concept. I like it.

Since I'm not that interested in wrestling vectors for days, and I started with a non-commercial font in the first place, I reached out to Fiverrrrr to see if anyone would do the running series title graphic for me. Next step is taking it to my cover artists and asking for the "clean up my cover" service. They might be willing to do the font work, but they also might cheapen out, and it is a major design element for me. The rest of it is a lot of tweaking font fills and kerning (whatever the hell kerning is. I joke. I do know what kerning is). And that sort of thing, they can do just fine.

I'll see. Right now Fiverr is being Fiverr. People missing the point as they frantically run their AI translation software in the background, hoping that being able to pretend to share a language is all they need to pretend to share background and aesthetics as well...and relentlessly hoping to upsell me into an entire cover-and-book-edit while they are at it.

***

ComfyUI's implementation of the WAN2.1 also ran, but illustrated a very basic problem with the current paradigm of AI. No, the belief of legions of middle-management to the contrary, plain-language is not the appropriate tool to communicate technical needs.

We do have a vocabulary of shots and cinematography but that isn't how shots, at the level of a film, are constructed and communicated. There's story-boards, for one. And, yes, there is something like that moving in.

Look, better analogy. I've worked around a bunch of choreographers. They will often demonstrate moves, and work through them and count them out while teaching them. They don't just say, "It's just a jump to the left, then a step to the right."

So shoving 250 words into a text prompt is not going to create the combination of actor move and camera move that the eye of an artist desires. It can only make a shot similar to the shots it has seen other people do, that might look good enough to convince someone without those skills that it is communicating the directorial intent.

Wednesday, September 3, 2025

Is nano banana qwen?

Actual title of a post on the StableDiffusion subreddit.

My Automatic1111 install croaked again. "Of course!" say the power users on the SD forums. "That's ancient. They stopped supporting it in mid-2024!"

That's how fast AI is moving. Yes, people are getting water-cooled Nvidia 5060 rigs set up (under an 16-core Ryzen or better). ComfyAI is old hat, with several other forks of Automatic1111, other independent run environments, and a front end for Comfy (which itself is basically a front end) that's got six different versions out there already.

Yeah, so I installed Comfy, even if it, too, is already old hat. But since another thing that is old hat is doing a clean install of Automatic1111 I just did that, too. Maybe this time I'll be smart and clone a backup (without the models, which is what takes up all the space).

It is old but I more or less know how to use it. It also is apparently much superior to ComfyUI for inpainting, and the thing I'm having fun with (I can't really call it workflow, since -- cue the Jurassic Park quote -- I'm not actually working on anything) is a process that is basically hyper-inpainting.

On a line with "assisted by software" towards one end and "created by the machine" on the other, I take some minor comfort that I'm leaning closer to the former than you might expect from the phrase "using AI." But after all, that's what editing is now over in book land; you are using AI tools, but they identify possible grammatical errors, and if you let them, make suggestions. You chose what to do in response. And "letting the AI change it" isn't the most common response.

This doesn't take away from the basic problems of AI. The morality (and possible legality) of the training data. The dangers of it, and not just misuse. And lurking under all of that, the black box nature of the thing. You are never quite in control and you never quite know what you are going to get, and for all that I could go on about serendipity (Rodin knocking a bust off the table and turning it into Man with a Broken Nose) and the ornery nature of organic tools (you never have total control over where that ink will go or how that paint will look)...

***

Revisited the cover concept again with a hybrid of Indiana Jones map and a bit of the Rivers of London look. That's basically a map, with the Beau Geste dotted line, the series running header as the dominant text element in swooping "adventure" font, and a compass rose with book number in the center.


Meanwhile writing is going slowly. I did a thousand words on the next scene yesterday. But even though it works, I'd prefer something with more conflict in it. So complete rewrite of that one.

I have a lot of Maytag Repairman time at work (we did an install yesterday; out of the building and something constructive to do, yay!) I can't concentrate well enough to write, so been reading up at (and on) Atomic Rockets.

Turns out Project Pluto wasn't even the height of insanity. Try Nuclear Salt Water rocket engine. That thing scared Scott Manley! And then there's what I can only describe as a nuclear solid rocket booster, for which one of the more interesting limitations is they have to make them really long and thin because there is no way of slowing down the China Syndrome melt-down which is heating up the lithium hydroxide reaction mass! 

I've also decided to not worry about lost opportunities for extra conflict and drama. The Athena Fox stories are more of a free-form exploration of a place, culture and time, where the solution of the mystery arises organically from the environment.

A plot which is intensively "Fetch Quest" oriented? Maybe The Tiki Stars is a good place to try that out.


Monday, September 1, 2025

Buy a bago, win-a-bago

You win a few, you lose a few.

I was looking up and down 54 and thinking about ghost towns. There's a cool one close to Orogrande (which itself is pretty weird). I got tempted by thoughts of abandoned silver mining towns, by Old West associations, or even by graveyards (cue "Ecstasy of Gold.")

But in the end, I didn't need to do a whole scene about Penny turning around the truck (and sending a text to Lon). I can just pick up with her meeting.

So I move on to thinking about my Alamogordo sources, but it felt a little too grand tour to send her to every notable thing in that town. And then I realized I could just set this upcoming scene at the burger joint. That hits a couple of useful plot bits; this changes Penny's relationship with "Dynel," and means the sheriff can say "You were seen in his company" when he goes asking around. And I do love a hub location. I sent most of the cast to the cafe in the last book, enough to make it a running gag.

Besides, I'm saving the giant pistachio.


It ain't all winning, though. At the same time, I realized I really need to go Oop North to interview this one guy. It wasn't right to leave Native Americans out of the story of White Sands, or the atomic age in New Mexico, so I had that in the story. And now that I've involved NAGPRA and have several key native characters, I have to do it. I have to send the story to the rez.

That does give me more freedom to focus in. To basically drop any cowboy stuff (another book, right). 

Largely, this is still moving to third-act territory, where all the basic settings and the cast are all here, and the reader won't have to go on learning new things. It makes sense to focus in.

I was also increasingly feeling I needed to focus in on New Mexico, and Egtved Girl didn't make the cut. I decided I was happy losing all of those vignettes except for bookend scenes; "White Sands Girl" (I still can't figure out which exact numbered trackway she was, but she was the one moving quick across mud, carrying a child, and encountering both a columbian mammoth and a Hanson's ground sloth. Busy day.)


And Valentina. Who encountered a whole hell of a lot of nothing (more nothing than most of us will ever experience) but had an even busier day.

But at the same time, the museum scene had been evolving (I haven't written it yet, even though it happens quite a bit further back than where I currently am). And I thought of a cool thing I could do that would both pull the threads together, and fill out the necessary space for Penny's epic hike through White Sands from I-haven't-decided-yet to the Trinity Site.

So...Lozen out. Lozen in. And Egtved Girl might still make the cut. But...the way this is being done, it takes a lot of the pressure I was feeling about having to do enough research to pull off 1-2,000 words on "woman walking" at various random points of history.

In any case, I'm pretty much plotted through to the end now. I've stopped updating the outline, but I have place-holder chapters out through to the end of the current act. Now I just need to find the focus to write (the hot weather is not helping!)

Friday, August 29, 2025

They call the wind Katrina

Where do ideas come from?

The problem in explaining that to the younger writer (young in the sense of experience) is that ideas don't arrive in words. They are working with, "I have this idea of a boy who discovers he is half-alien and inherited a laser sword from his real father and then has to fight an evil empire."

When you've been in the writing business for a while, they are rather more inchoate. Stuff is sort of getting caught in the baleen and at some random point you turn around and realize you could connect this to this ...and that's when it becomes something you can put down in your notes.


That's where I was when the idea of doing a "waking up with amnesia" story crossed paths with "boy, I'm tired of First Person." And then it bumped into "I'd love to do something with New Orleans, especially in re Katrina and the diaspora."

And that's when that town and that historical context started showing up everywhere. Someone in a writing subreddit was asking about researching the Big Easy for an urban fantasy, and wanted to set it around 2005. KPFA has been doing a mammoth retrospective (turns out there is a series I wish was on my streaming resource).

I have a mild connection due to working a show that celebrated the voices of the diaspora, bringing many transplanted musicians from New Orleans together for a concert.

Those ideas of nucleation and growth are strong; when that parent crystal is big enough, it jars all the other stuff around it like a Flyer in Conway's Life. It attracts things that you didn't realize could fit, and turns them into ever more elaborate structures. And it just sort of happens, without having to write an outline or a story bible or pay up your licenses on World Anvil.


That's how ideas get out of hand. It isn't a matter of hammering away, it isn't a blinding flash of inspiration, it isn't a writing prompt you could ask CatGTP to spit out. It is just how the writer's mind works.

Because it really is the same stuff you are doing all the time anyhow. All the time, you are fleshing out, adding details, discovering or forging connections. Not just for the big overall plot, not just at the outline level. Right down to individual lines of dialogue.

Well...back to draft #3 of Penny getting checked out for possible radiological contamination. I've been weak, and injured my back as well, so not the most productive week. But I did rework part of the first big Mary Cartwright chapter.

Her personality was inspired by a Tewa writer. But I just discovered and archive-binged a blog from an equally angry Cherokee writer, Native Appropriations. Which made me rethink a couple of things Mary said, especially re Cody and his revisionist western series about a young boy named Chato.

Plus I'm trying to get the opposition into the picture as an active force. I've added a mystery truck to that chapter. Not sure how I can do anything earlier, and my head canon says Evil Kitty didn't start worrying about Penny until after they found out through their military contacts that she'd been messing around with Site Theta.


And....I just discovered there is an abandoned Six Flags in New Orleans. Trashed by Katrina and never rebuilt. How could you not use something like that in a story featuring a student archaeologist turned urban explorer?

Sunday, August 24, 2025

Cross-contamination

Some writers, more likely the new writers than the experienced, avoid books when they are actively writing. They are afraid that something will rub off.

I didn't used to 

have that fear. I wonder, though. Writing in narrative voice can be difficult at times. First Person is practically defined by having a distinctive voice; it is Penny's verbal ticks and manners that hit me worst out of all the travails of writing in First Person. (Seconded by her habit of covering up what she really feels.)

And I don't have a story bible for it. I've always sort of thought that was a thing you did when you created a distinctive character voice. This one does not use contractions. This one is enamored of sesquipedalian verbiage. I'm always reminded in this context of a bit in Connecticut Yankee when Twain's protagonist recommends giving each character in Sandy's rambling chivalric tale be giving a different accent or verbal tic. Cue Sandy inserting "begorrah" after every single line of the Irish knight.

For Penny, I just write the stuff. I go into her head and I jot down what comes out. And two times out of three, what she chose to say didn't work for the story or the audience or was just too weird and I have to go back and delete it. It is all instinct. I've nothing written down. 

So, yeah, I got suspicious when I archive-binged the first season of Wednesday and noticed that Penny seemed to be getting more snarky than usual.

Alexis Carew clinched it for me. Penny keeps wanting to say things like "I would admire if were it done promptly" and I have to stop her. Instanter.


I've got her getting a blood draw at the FOMC at Holloman AFB, part of the 49th Medical Group. "Perseverance Prevails," their motto, and I'm having to hold Penny back from muttering something about how many sea anemone are enemies of her. Oddly enough, I tried my local medical supply guy on it and he pronounced it no problem.

No, I didn't ask him about my medical questions. He doesn't do that. He'll sell you a wheelchair. He doesn't do medicine.

I was able to pull up enough data to figure out what kinds of tests these guys could be running on Penny to ward off what she calls her potential to turn into her own nightlight. She means glowing in the dark. And so far nobody has admired her clever turn of phrase (she keeps hoping). It also works out very well for my purposes, because half of it can be done by any lab, and the other half gets mailed off to the CDC regardless of who does it.

I do wish I has some internal photographs or video of the place. But more than that, what kind of procedures they might have, how they might act. Ah, well. I've pretty much given up on strict accuracy. If someone reads the book on the flightline and sends in an angry review -- it will be one more review than I ever expect to get. For this book or for the last, either.

Also, Reacher did it for me at last on accuracy. Dan Brown was bad enough. Lee Child makes fewer mistakes, but a big part of the books is how Reacher is always thinking and doing the math and working out just how far the bad guy drove or why the cafe has to be part of a money-laundering operation.

And its wrong. Weirdly wrong. I mean...there are a hell of a lot of guns in these books and I swear that at least one of the pistols was loaded with shells. There might have even been a clip...