đ Executive Summary
TL;DR: AI-generated headshots often look artificial due to model oversmoothing; achieve realism by employing advanced prompt engineering with specific positive and negative prompts, fine-tuning models via LoRA with personal photos, and utilizing hybrid workflows for final touches.
đŻ Key Takeaways
- Counteracting Model Oversmoothing: Explicitly include âskin textureâ and âporesâ in positive prompts and use weighted negative prompts like â(plastic, doll, smooth skin, airbrushed:1.3)â to combat the generic, smoothed-out appearance from base diffusion models.
- Personalized Realism with LoRA: Train a Low-Rank Adaptation (LoRA) model using 15-20 varied personal photos to embed specific facial structures into the AI, enabling consistent and highly realistic generations that generic prompts cannot achieve.
- Optimizing Sampler Settings: Fine-tune CFG Scale (e.g., 6.5 for naturalness) and Denoising Strength (e.g., 0.4-0.6 for img2img) to control the AIâs adherence to the prompt and preserve subtle textures, preventing the âdenoising trapâ that removes realism.
Tired of AI-generated headshots looking like plastic dolls? Learn how to inject realism by fine-tuning models with your own photos (LoRA) and mastering advanced prompting techniques for truly lifelike results.
I See Your AI Headshot and Raise You One Uncanny Valley: Escaping the Wax Museum
I still remember the Slack message from Marketing. âHey Darian, can you spin up that âAI thingâ and get us new headshots for the exec team? We need them for the new âAbout Usâ page by Friday.â Simple enough, I thought. I grabbed a few of the CEOâs approved photos from our press kit, fired up a Stable Diffusion instance, and ran a basic prompt. The result? A perfectly coiffed, wrinkle-free, soulless mannequin that looked like our CEO had a horrifying accident at a candle-making factory. It was technically him, but it had zero life. Thatâs when I knew we had to go deeper than just a simple text prompt.
So, Why Does This Keep Happening? The Root of the Plastic Look
Before we jump into the fixes, you need to understand why this happens. Itâs not just âbad AI.â Itâs a combination of a few things:
- Model Oversmoothing: Base diffusion models are trained on millions of images. To create a âgeneralâ human face, they average out features. This process naturally smooths out imperfectionsâpores, fine lines, asymmetrical featuresâwhich are the very things that make a face look real.
- The Denoising Trap: The core process of diffusion is starting with noise and âdenoisingâ it into an image based on your prompt. If the denoising strength is too high, it wipes away subtle textures, resulting in that glossy, airbrushed finish.
- Lack of Specificity: A prompt like âphoto of a man in a suitâ is asking the model to pull from its vast, generic knowledge. It doesnât know *your* face, it knows the *idea* of a face.
The goal isnât to just make a picture; itâs to guide the model away from its generic defaults and force it to add back the chaos of reality. Hereâs how we do it in the trenches.
Solution 1: The Quick Fix (Prompt Engineering & Sampler Fu)
This is your first line of defense. Itâs fast, requires no custom models, and can often get you 80% of the way there. Itâs all about telling the model what to add and, more importantly, what to avoid.
The Prompt Breakdown
Instead of a simple prompt, we get hyper-specific and use negative prompts to fight the âwax museumâ effect. Letâs assume youâre using a tool like Automatic1111 or any other standard UI.
Positive Prompt:
(photograph of a 40-year-old man), professional headshot, sharp focus, (skin texture:1.2), pores, detailed skin, slight smile, soft studio lighting, Canon EOS 5D Mark IV, 85mm f/1.8 lens
Negative Prompt:
(plastic, doll, cartoon, 3d, render:1.3), painting, art, blurry, oversaturated, smooth skin, airbrushed, unnatural
Notice the key elements: Weâre explicitly asking for âskin textureâ and âporesâ and using weights (word:1.2) to increase their importance. In the negative prompt, weâre actively telling it to avoid the things that make images look fake. This is the equivalent of putting guardrails on the generation process.
Tweak Your Sampler Settings
Donât just hit âGenerate.â Play with these two settings:
| Setting | What it Does & Why You Should Care |
| CFG Scale | How strictly the AI follows your prompt. A low value (e.g., 4-6) gives it more creative freedom, which can feel more natural. A high value (e.g., 7-10) sticks closer to your prompt but can feel rigid. Start around 6.5. |
| Denoising Strength (for img2img) | If youâre using an existing photo as a base (img2img), this is critical. A value of 0.4-0.6 will change the style and lighting while preserving the facial structure. Go higher, and you start getting a different person. |
Solution 2: The Permanent Fix (Train a LoRA On Yourself)
This is the real deal. When prompt engineering isnât enough, you need to teach the model exactly what you look like. We do this by training a LoRA (Low-Rank Adaptation). Think of it as a small, lightweight âpluginâ for the main model that contains information about a specific person or style.
This is where my DevOps hat comes on. You donât need a massive GPU farm for this. You can rent a GPU instance (like an A100 on GCP or AWS) for an hour or use a cloud service that automates it.
The High-Level Workflow:
- Gather Your Data: Collect 15-20 high-quality, varied photos of yourself. Different angles, different lighting, different expressions. No sunglasses, no hats.
-
Prepare & Tag: Use a tool (like the automated taggers in Kohya_ss) to caption each image. This tells the training process whatâs in the picture (e.g., âa photo of dvance_manâ). The unique keyword
dvance\_manis what youâll use to call yourself in the prompt. -
Train the LoRA: Load your images and captions into a training UI like Kohya_ss. Youâll set parameters like learning rate and number of epochs. This process usually takes 20-40 minutes on a decent GPU. It spits out a small file (e.g.,
dvance\_man.safetensors, maybe 100MB). - Generate with Your LoRA: Now, you use a special prompt that includes your LoRA.
Positive Prompt:
professional headshot of <lora:dvance_man:0.8>, sharp focus, detailed skin, corporate office background
// The <lora:lora_name:weight> syntax tells the generator to apply your trained LoRA.
// The weight (0.8 here) controls its intensity.
The result is night and day. Because the model now has specific data about your facial structure, it can generate you realistically in any context you ask for. This is how you get consistency and realism that prompting alone canât achieve.
A Word of Warning: Be mindful of where you upload your photos. Using a local setup or a trusted, private cloud instance (like a Jupyter Notebook on Vertex AI) gives you full control. Iâm hesitant to use free online services where I donât know how my training data is being stored or used.
Solution 3: The âHybridâ Option (AI Base, Photoshop Finish)
Sometimes you get an image thatâs 95% perfect. The composition is great, the lighting is on point, but the eyes look a little⌠off. Or the skin is just a bit too smooth. This is when you stop fighting the model and bring in other tools.
This is a hacky but incredibly effective workflow weâve used for one-off images:
- Generate the Base: Use Solution 1 or 2 to get a strong base image. The overall structure should be good.
- Inpaint the Problem Areas: In your image generatorâs UI, use the âInpaintâ feature. Mask just the skin on the face, leaving the eyes, hair, and clothes alone.
- Run a Detail Pass: Use a prompt like âultra-detailed skin texture, pores, imperfectionsâ with a very low denoising strength (0.2-0.35). The AI will only re-generate the masked area, adding the texture you asked for without changing the face.
- Final Touches Elsewhere: If that fails, donât be afraid to take the image into Photoshop, GIMP, or Krita. Use frequency separation to add texture, or use the liquify tool to fix a slightly wonky eye. Combining the creative power of the AI with the precision of manual editing is often the fastest path to a perfect result.
At the end of the day, these tools are just thatâtools. Getting a great result isnât about finding a magic prompt. Itâs about understanding the âwhyâ behind the waxy faces and using a combination of prompting, data, and workflow to force the machine to bend to reality, not the other way around.
đ Read the original article on TechResolve.blog
â Support my work
If this article helped you, you can buy me a coffee:

Top comments (0)