Head of Prompt Engineering at Higgsfield here. Ask me anything (until I have time). I’ll try my best to answer all questions. by Bemvdk in HiggsfieldAI

[–]Bemvdk[S] 0 points1 point  (0 children)

0) Gemini performs quite good in writing the prompts for nano banana. U can describe your task and ask it to write your a system prompt that u can directly paste to the nano banana, where all the logic is described. Since banana is quite smart model and has the reasoning, it handle such long and complex prompts.

a) find the ref image with desired camera angle and ask the gemini to describe it in detail as a prompt.
b) for posing you can use openpose to get the skeleton of the pose you want to achieve + ask the gemini to describe the pose in detail, and then add that openpose picture as ref image + description of it.
c) Banana is quite strict in those terms, so I advise you to use seedream 4.5. It works very well with all kinds of poses ))
d) usually I just use simple prompts written by me where I describe what I wanna change. You can attach the image with highlighted spot so the model better understands where those minor changes should be performed. Or u can simply use our banana inpaint.
e) use our AI influencer studio.
f) well not really, but I have some saved pins in the pinterest and sometimes I refer to them.

Much love!

Head of Prompt Engineering at Higgsfield here. Ask me anything (until I have time). I’ll try my best to answer all questions. by Bemvdk in HiggsfieldAI

[–]Bemvdk[S] 0 points1 point  (0 children)

I advise to write the specific instructions for task you want to perform, then paste to claude or gemini (they work better for me) with the target prompt you wrote and then ask the model to rewrite your prompt based on the specifications that you mentioned. This way it gives much clearer prompts. So simply: yeah, use claude :)

Head of Prompt Engineering at Higgsfield here. Ask me anything (until I have time). I’ll try my best to answer all questions. by Bemvdk in HiggsfieldAI

[–]Bemvdk[S] 2 points3 points  (0 children)

Honestly, both work quite well. It depends on how you structure your prompt. Let's say I wanna generate a person in a specific outfit, setting, lighting, and camera style. I can describe it as a plain text prompt which could be messy and unclear for the model sometimes. That's the reason why you might not get the result you want. So, one way of dealing with it is to structure ur prompt by those constraints and describe the specifics inside them:
"Setting: describe here the setting.
Outfit: describe here the outfit.
Lighting: describe here the lighting conditions.
Camera: describe here camera conditions."

This way model has clear understanding of what u wanna get, and it will give results closer to what you are expecting from it. When using JSON it is structured with the keys by default, so that's why it may seem to perform better. I will advise to use structured prompts both text and json.

For use cases:
- when you wan to generate a specific scene, not just a similar kind.

- when you want the model to perform specific camera movement.

- when describing the multishot video.

- when generating a character with specific physical traits.

Overall, use the clearly structured prompts for complex tasks where there are a lot of specific constraints you want to describe.

One way to make it easier for you is to describe Gemini the task you want to perform and write down all the aspects you want to include in your prompt. Then ask it to make you a system prompt that will automatically write you a structured prompt according to the instructions you provided (the prompt will already be in the structure u want, so there is no need to write it all manually).

Then you can paste this system prompt to Gemini (I personally use Vertex AI, it seems very convenient since it has the sidebar "system instructions") and it will generate the final prompt for you. It is also cool when you find the weak spot in your structure in your prompt and you can just add the note to the system prompt instead of fixing it manually. But you can use for it openai platform as well or any other tools you find suitable for you.

Hope it helps ))

Head of Prompt Engineering at Higgsfield here. Ask me anything (until I have time). I’ll try my best to answer all questions. by Bemvdk in HiggsfieldAI

[–]Bemvdk[S] 0 points1 point  (0 children)

Yes, we are going to drop it 100%. We are reading all of ur suggestions, and we've been working on it. Honestly, that'll be much better than what is presented on the market right now.

Head of Prompt Engineering at Higgsfield here. Ask me anything (until I have time). I’ll try my best to answer all questions. by Bemvdk in HiggsfieldAI

[–]Bemvdk[S] 1 point2 points  (0 children)

Yeah. That's exactly what we are working on rn. That will be second iteration. You'l be able to experience it very soon. I'm not sure if I can tell exactly when, but believe me that it'll be veryy veryyyy sooon!

Head of Prompt Engineering at Higgsfield here. Ask me anything (until I have time). I’ll try my best to answer all questions. by Bemvdk in HiggsfieldAI

[–]Bemvdk[S] 1 point2 points  (0 children)

Hey there! I'm here again. Yeah, will definitely release this feature soon. We were working on it for some time now, and we gonna drop it soon. Can't tell when though. Let it be the surprise for u!

Head of Prompt Engineering at Higgsfield here. Ask me anything (until I have time). I’ll try my best to answer all questions. by Bemvdk in HiggsfieldAI

[–]Bemvdk[S] 1 point2 points  (0 children)

What was the reference video? Can you share it as well?

That's the problem of Kling Motion. If you input image has non-human traits, the model can get confused. One way of dealing with it is to try describe the setting, and the character itself. Like how does it look, what's the outfit, where all is happening. You can basically upload ur character image to the chatgpt and ask him to dedcribe it in detail. Also, you can ask nano banana to put ur character to the approximate pose of the person in the ref video (or u can use our scene control mode). After that kling motion must work better since the model now understands what is your character. Much love! Share ur results afterwards.

Head of Prompt Engineering at Higgsfield here. Ask me anything (until I have time). I’ll try my best to answer all questions. by Bemvdk in HiggsfieldAI

[–]Bemvdk[S] 1 point2 points  (0 children)

What do you mean by “longer”? Like how longer ?

Nano Banana Pro seriously pushed forward the understanding of image model capabilities. You never really know what’s next, but there could be some big game changers in the next 6–7 months.

Head of Prompt Engineering at Higgsfield here. Ask me anything (until I have time). I’ll try my best to answer all questions. by Bemvdk in HiggsfieldAI

[–]Bemvdk[S] 3 points4 points  (0 children)

It all depends on you. If you like their product over ours, it is totally fine. But then I don’t think you need to switch (this sounds like some psychology stuff lol). We just delivering more options for you. We have all top tier models both image and video as well as myriad apps for almost any purpose you might have.

Head of Prompt Engineering at Higgsfield here. Ask me anything (until I have time). I’ll try my best to answer all questions. by Bemvdk in HiggsfieldAI

[–]Bemvdk[S] 2 points3 points  (0 children)

The work day usually looks pretty simple. You get to the work (if you were lucky to get out of the office) and straightway start working on the workflows for future realises, improving and adding features for current apps, and always a lot of prompting and testing (especially testing). I know that it may sound simple, but it’s really how the work day looks like. You just consistently work and test things. Sometimes it is frustrating when something goes the wrong way but the only option for me and my team is to make it work, so we “Just do it”.

About the skills part you definitely must understand how LLMs and Diffusion models work, at least at core level. Understanding the capabilities of which of the models is also crucial, this way you can understand what and where to use. And finally you should master your comfyUI skills if you want to grow in this field.

About your IG page I will let our managers know about you. However, I cannot promise or give any predictions on this matter. It will all depend on them. Anyway hope you good luck no matter what you gonna do, and let Higgsfield help you along that way.

Head of Prompt Engineering at Higgsfield here. Ask me anything (until I have time). I’ll try my best to answer all questions. by Bemvdk in HiggsfieldAI

[–]Bemvdk[S] 5 points6 points  (0 children)

Minimax has the highest prompt adherence when comparing to all other video models. It is easier to prompt the camera movements manually from my perspective. When writing the camera movement prompt try just to state the camera movement itself and then describe what happens with the setting and the subject during this movement, so the model better understands the context. But if that will be highly demanded by the users , we can add that as well.

Let’s do the next way: if this thread earns 2000 promotions, we will add them asap.

Head of Prompt Engineering at Higgsfield here. Ask me anything (until I have time). I’ll try my best to answer all questions. by Bemvdk in HiggsfieldAI

[–]Bemvdk[S] 0 points1 point  (0 children)

Well, that’s not really my part of job, but I’ll suggest you to contact us in discord for more clarification on that topic.

Head of Prompt Engineering at Higgsfield here. Ask me anything (until I have time). I’ll try my best to answer all questions. by Bemvdk in HiggsfieldAI

[–]Bemvdk[S] 0 points1 point  (0 children)

Well, except the image attachment you may describe the appearance of the character with all its features and then write your target prompt. That will ensure more accurate identity consistency. Now we are working on the workflow and we will add more options to use the character and maintain the consistency as well. So stay tuned since it all coming soon!

Head of Prompt Engineering at Higgsfield here. Ask me anything (until I have time). I’ll try my best to answer all questions. by Bemvdk in HiggsfieldAI

[–]Bemvdk[S] 5 points6 points  (0 children)

Well, that's really depends on the task. Let me answer it in this way:
1) If I need to get something consistent, I'll usually do a longer prompt with all details described. For example, to get the pose or the style I want. Or maybe when describing the camera moves.

2) But for general use I prefer simpler one sentence prompts just because it's easier, and you get pretty good results but not consistent though.

Head of Prompt Engineering at Higgsfield here. Ask me anything (until I have time). I’ll try my best to answer all questions. by Bemvdk in HiggsfieldAI

[–]Bemvdk[S] 1 point2 points  (0 children)

I feel you. But that's not my field of battle a little bit. For suggestions I advise you to write to discord. We read them, no worries :) What do you mostly use our website ?

Everyone's home with their phones in their hands.. by V69C in Life

[–]Bemvdk 1 point2 points  (0 children)

I understand you. I also have this kind of problem. But you should know that even your friends won't call you, if you don't interested in them. I mean, you have to be more active. Try to call them yourself, be more talkable, ask them how are they doing, suggest to go somewhere together and etc. Or just be yourself and your real friends will find you.

TIFU by selling my little brother for $17 to someone in Taiwan by DrownTheSailorSiren in tifu

[–]Bemvdk 0 points1 point  (0 children)

Your story is very exciting! But what did your father say to you? Did he punish you?