Initial implementation of Stable Diffusion
- Needs to add attachment to discord - Needs to delete image afterward - Needs to calculate token usage
This commit is contained in:
263
utilities/stabilityHelpers.js
Normal file
263
utilities/stabilityHelpers.js
Normal file
@@ -0,0 +1,263 @@
|
||||
// Debug
|
||||
const { DebugBuilder } = require("../utilities/debugBuilder");
|
||||
const log = new DebugBuilder("server", "stabilityController");
|
||||
|
||||
// Modules
|
||||
const Generation = require("../controllers/stabilityAi/generation_pb");
|
||||
const { GenerationServiceClient } = require("../controllers/stabilityAi/generation_pb_service");
|
||||
const fs = require("fs");
|
||||
|
||||
const { grpc } = require("@improbable-eng/grpc-web");
|
||||
const GRPCWeb = grpc;
|
||||
|
||||
const path = require("node:path");
|
||||
|
||||
function isImageArtifact(artifact) {
|
||||
log.DEBUG("Checking if image is an artifact", artifact);
|
||||
if (artifact.getType() !== Generation.ArtifactType.ARTIFACT_IMAGE) return false;
|
||||
if (artifact.getFinishReason() !== Generation.FinishReason.NULL) return false;
|
||||
return artifact.hasBinary();
|
||||
}
|
||||
exports.isImageArtifact = isImageArtifact;
|
||||
|
||||
function isNSFWFilteredArtifact(artifact) {
|
||||
if (artifact.getType() !== Generation.ArtifactType.ARTIFACT_IMAGE) return false;
|
||||
if (artifact.getFinishReason() !== Generation.FinishReason.FILTER) return false;
|
||||
return true;
|
||||
}
|
||||
exports.isNSFWFilteredArtifact = isNSFWFilteredArtifact;
|
||||
|
||||
/** Builds a generation request for a specified engine with the specified parameters. */
|
||||
function buildGenerationRequest(engineID, params) {
|
||||
if (params.type === "upscaling") {
|
||||
const request = new Generation.Request()
|
||||
request.setEngineId(engineID)
|
||||
request.setRequestedType(Generation.ArtifactType.ARTIFACT_IMAGE)
|
||||
request.setClassifier(new Generation.ClassifierParameters())
|
||||
|
||||
const imageParams = new Generation.ImageParameters()
|
||||
if ("width" in params && !!params.width) {
|
||||
imageParams.setWidth(params.width)
|
||||
} else if ("height" in params && !!params.height) {
|
||||
imageParams.setHeight(params.height)
|
||||
}
|
||||
request.setImage(imageParams)
|
||||
request.addPrompt(createInitImagePrompt(params.initImage))
|
||||
|
||||
return request
|
||||
}
|
||||
|
||||
const imageParams = new Generation.ImageParameters()
|
||||
if (params.type === "text-to-image") {
|
||||
params.width && imageParams.setWidth(params.width)
|
||||
params.height && imageParams.setHeight(params.height)
|
||||
}
|
||||
|
||||
// Set the number of images to generate (Default 1)
|
||||
params.samples && imageParams.setSamples(params.samples)
|
||||
|
||||
// Set the steps (Default 30)
|
||||
// Represents the amount of inference steps performed on image generation.
|
||||
params.steps && imageParams.setSteps(params.steps)
|
||||
|
||||
// Set the seed (Default 0)
|
||||
// Including a seed will cause the results to be deterministic.
|
||||
// Omitting the seed or setting it to `0` will do the opposite.
|
||||
params.seed && imageParams.addSeed(params.seed)
|
||||
|
||||
// Set the sampler (Default 'automatic')
|
||||
// Omitting this value enables 'automatic' mode where we choose the best sampler for you based
|
||||
// on the current payload. For example, since CLIP guidance only works on ancestral samplers,
|
||||
// when CLIP guidance is enabled, we will automatically choose an ancestral sampler for you.
|
||||
if (params.sampler) {
|
||||
const transformType = new Generation.TransformType()
|
||||
transformType.setDiffusion(params.sampler)
|
||||
imageParams.setTransform(transformType)
|
||||
}
|
||||
|
||||
// Set the Engine
|
||||
// At the time of writing, valid engines are:
|
||||
// stable-diffusion-v1,
|
||||
// stable-diffusion-v1-5
|
||||
// stable-diffusion-512-v2-0
|
||||
// stable-diffusion-768-v2-0
|
||||
// stable-diffusion-512-v2-1
|
||||
// stable-diffusion-768-v2-1
|
||||
// stable-inpainting-v1-0
|
||||
// stable-inpainting-512-v2-0
|
||||
// esrgan-v1-x2plus
|
||||
const request = new Generation.Request()
|
||||
request.setEngineId(engineID)
|
||||
request.setRequestedType(Generation.ArtifactType.ARTIFACT_IMAGE)
|
||||
request.setClassifier(new Generation.ClassifierParameters())
|
||||
|
||||
// Set the CFG scale (Default 7)
|
||||
// Influences how strongly your generation is guided to match your prompt. Higher values match closer.
|
||||
const samplerParams = new Generation.SamplerParameters()
|
||||
params.cfgScale && samplerParams.setCfgScale(params.cfgScale)
|
||||
|
||||
const stepParams = new Generation.StepParameter()
|
||||
stepParams.setScaledStep(0)
|
||||
stepParams.setSampler(samplerParams)
|
||||
|
||||
const scheduleParams = new Generation.ScheduleParameters()
|
||||
if (params.type === "image-to-image") {
|
||||
// If we're doing image-to-image generation then we need to configure
|
||||
// how much influence the initial image has on the diffusion process
|
||||
scheduleParams.setStart(params.stepScheduleStart)
|
||||
if (params.stepScheduleEnd) {
|
||||
scheduleParams.setEnd(params.stepScheduleEnd)
|
||||
}
|
||||
} else if (params.type === "image-to-image-masking") {
|
||||
// Step schedule start is always 1 for masking requests
|
||||
scheduleParams.setStart(1)
|
||||
}
|
||||
|
||||
stepParams.setSchedule(scheduleParams)
|
||||
|
||||
// Set CLIP Guidance (Default: None)
|
||||
// NOTE: This only works with ancestral samplers. Omitting the sampler parameter above will ensure
|
||||
// that we automatically choose an ancestral sampler for you when CLIP guidance is enabled.
|
||||
if (params.clipGuidancePreset) {
|
||||
const guidanceParameters = new Generation.GuidanceParameters()
|
||||
guidanceParameters.setGuidancePreset(params.clipGuidancePreset)
|
||||
stepParams.setGuidance(guidanceParameters)
|
||||
}
|
||||
|
||||
imageParams.addParameters(stepParams)
|
||||
request.setImage(imageParams)
|
||||
|
||||
params.prompts.forEach(textPrompt => {
|
||||
const prompt = new Generation.Prompt()
|
||||
prompt.setText(textPrompt.text)
|
||||
|
||||
// If provided, set the prompt's weight (use negative values for negative weighting)
|
||||
if (textPrompt.weight) {
|
||||
const promptParameters = new Generation.PromptParameters()
|
||||
promptParameters.setWeight(textPrompt.weight)
|
||||
prompt.setParameters(promptParameters)
|
||||
}
|
||||
|
||||
request.addPrompt(prompt)
|
||||
})
|
||||
|
||||
// Add image prompts if we're doing some kind of image-to-image generation or upscaling
|
||||
if (params.type === "image-to-image") {
|
||||
request.addPrompt(createInitImagePrompt(params.initImage))
|
||||
} else if (params.type === "image-to-image-masking") {
|
||||
request.addPrompt(createInitImagePrompt(params.initImage))
|
||||
request.addPrompt(createMaskImagePrompt(params.maskImage))
|
||||
}
|
||||
|
||||
return request
|
||||
}
|
||||
exports.buildGenerationRequest = buildGenerationRequest;
|
||||
|
||||
function createInitImagePrompt(imageBinary) {
|
||||
const initImageArtifact = new Generation.Artifact()
|
||||
initImageArtifact.setBinary(imageBinary)
|
||||
initImageArtifact.setType(Generation.ArtifactType.ARTIFACT_IMAGE)
|
||||
|
||||
const initImageParameters = new Generation.PromptParameters()
|
||||
initImageParameters.setInit(true)
|
||||
|
||||
const initImagePrompt = new Generation.Prompt()
|
||||
initImagePrompt.setParameters(initImageParameters)
|
||||
initImagePrompt.setArtifact(initImageArtifact)
|
||||
|
||||
return initImagePrompt
|
||||
}
|
||||
|
||||
function createMaskImagePrompt(imageBinary) {
|
||||
const maskImageArtifact = new Generation.Artifact()
|
||||
maskImageArtifact.setBinary(imageBinary)
|
||||
maskImageArtifact.setType(Generation.ArtifactType.ARTIFACT_MASK)
|
||||
|
||||
const maskImagePrompt = new Generation.Prompt()
|
||||
maskImagePrompt.setArtifact(maskImageArtifact)
|
||||
|
||||
return maskImagePrompt
|
||||
}
|
||||
|
||||
/** Executes a GenerationRequest, abstracting the gRPC streaming result behind a Promise */
|
||||
async function executeGenerationRequest(
|
||||
generationClient,
|
||||
request,
|
||||
metadata
|
||||
) {
|
||||
try {
|
||||
const stream = generationClient.generate(request, metadata)
|
||||
const answers = await new Promise((resolve, reject) => {
|
||||
const answers = new Array()
|
||||
|
||||
stream.on("data", data => answers.push(data))
|
||||
stream.on("end", () => resolve(answers))
|
||||
stream.on("status", status => {
|
||||
if (status.code === 0) return
|
||||
reject(status.details)
|
||||
})
|
||||
})
|
||||
|
||||
return extractArtifacts(answers)
|
||||
} catch (err) {
|
||||
return err instanceof Error ? err : new Error(JSON.stringify(err))
|
||||
}
|
||||
}
|
||||
exports.executeGenerationRequest = executeGenerationRequest;
|
||||
|
||||
function extractArtifacts(answers) {
|
||||
const imageArtifacts = new Array()
|
||||
const filteredArtifacts = new Array()
|
||||
|
||||
for (const answer of answers) {
|
||||
for (const artifact of answer.getArtifactsList()) {
|
||||
if (isImageArtifact(artifact)) {
|
||||
imageArtifacts.push(artifact)
|
||||
} else if (isNSFWFilteredArtifact(artifact)) {
|
||||
filteredArtifacts.push(artifact)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
return { filteredArtifacts, imageArtifacts }
|
||||
}
|
||||
|
||||
/** Generation completion handler - replace this with your own logic */
|
||||
function onGenerationComplete(response) {
|
||||
if (response instanceof Error) {
|
||||
log.ERROR("Generation failed", response)
|
||||
throw response
|
||||
}
|
||||
|
||||
log.DEBUG(
|
||||
`${response.imageArtifacts.length} image${
|
||||
response.imageArtifacts.length > 1 ? "s" : ""
|
||||
} were successfully generated.`
|
||||
)
|
||||
|
||||
// Do something with NSFW filtered artifacts
|
||||
if (response.filteredArtifacts.length > 0) {
|
||||
log.DEBUG(
|
||||
`${response.filteredArtifacts.length} artifact` +
|
||||
`${response.filteredArtifacts.length > 1 ? "s" : ""}` +
|
||||
` were filtered by the NSFW classifier and need to be retried.`
|
||||
)
|
||||
}
|
||||
|
||||
// Do something with the successful image artifacts
|
||||
response.imageArtifacts.forEach(artifact => {
|
||||
try {
|
||||
const writePath =
|
||||
fs.writeFileSync(
|
||||
path.resolve(__dirname, `../.generations/image-${artifact.getSeed()}.png`),
|
||||
Buffer.from(artifact.getBinary_asU8())
|
||||
)
|
||||
} catch (error) {
|
||||
log.ERROR("Failed to write resulting image to disk", error)
|
||||
}
|
||||
})
|
||||
|
||||
// For browser implementations: you could use the `artifact.getBinary_asB64()` method to get a
|
||||
// base64 encoded string and then create a data URL from that and display it in an <img> tag.
|
||||
}
|
||||
exports.onGenerationComplete = onGenerationComplete;
|
||||
Reference in New Issue
Block a user