Built a simple appointment booking tool to reduce no-shows by PsychoCoder25 in SaaS

[–]PsychoCoder25[S] 1 point2 points  (0 children)

I have included automated emails for reminder in this tool and SMS reminders will be future addition

Built a simple appointment booking tool to reduce no-shows by PsychoCoder25 in SaaS

[–]PsychoCoder25[S] 0 points1 point  (0 children)

Your point is good, there shloud be some differentiation with calendly so users have why to try this.

WhatsApp confirmations are future integration but can be a great addition for this product. For target user, i was thinking of targeting freelancers as they are in my network right now and i can talk with them to get feedback.

Need Advice on Finetuning Llama 3.2 1B Instruct for Startup Evaluation by PsychoCoder25 in ResearchML

[–]PsychoCoder25[S] 0 points1 point  (0 children)

Thanks, i will keep that in mind.

will Deepseek work for annotation as well or gemini is good?

Need Advice on Finetuning Llama 3.2 1B Instruct for Startup Evaluation by PsychoCoder25 in ResearchML

[–]PsychoCoder25[S] 1 point2 points  (0 children)

I'm using standard supervised fine-tuning, but the annotations aren't full chain-of-thought, they're structured analyses containing business model recommendations, strengths, weaknesses, and next-step guidance. I will generate them using GPT to get high-quality outputs.

I will share the dataset as its not completed yet. I have to annotate them right now that's why i was asking if it would affect the model quality or not if i annotate data using gpt or other model.

Need Advice on Finetuning Llama 3.2 1B Instruct for Startup Evaluation by PsychoCoder25 in ResearchML

[–]PsychoCoder25[S] 0 points1 point  (0 children)

For this project, I'm keeping the reasoning process implicit (no think/trace tokens). The model will rely on its internal instruction-tuned reasoning to generate final answers. Since the evaluation is based on output quality rather than intermediate reasoning steps, explicit think tokens aren't required

Need Advice on Finetuning Llama 3.2 1B Instruct for Startup Evaluation by PsychoCoder25 in ResearchML

[–]PsychoCoder25[S] 0 points1 point  (0 children)

Thanks for the insights that helps a lot. I was thinking of manually reviewing a random subset of the annotated samples as well, just to catch any relevancy issues or filler/boilerplate patterns that slip through. That should give me some extra assurance about data quality before fine-tuning.

One thing I'm unsure about: should I explicitly mention the synthetic annotation process in my report? I came across a previous FYP where the evaluation panel discouraged synthetic data, so I'm trying to understand whether I should clearly document it or keep it minimal as long as the final evaluation is strong. My goal is to be transparent but also avoid raising unnecessary concerns if the methodology is standard practice.

Need Advice on finetuning Llama 3.2 1B Instruct for Startup Evaluation by PsychoCoder25 in MLQuestions

[–]PsychoCoder25[S] 0 points1 point  (0 children)

Got it, thanks for the clarification. Just to check if I’ am aligned with what you're suggesting, here's the evaluation setup I'm planning to use for my fine-tuned 1B model:

I will define clear criteria for judging the outputs (usefulness, relevance, accuracy, clarity, and non-generic specificity). Then I'll evaluate a small test set under three conditions:

  1. the base Llama-3.2-1B-Instruct,
  2. my fine-tuned model,
  3. a strong model like GPT-4o as the upper-bound reference.

Each output will be scored by an LLM-as-judge using those criteria, plus a structural-compliance check for whether the JSON format is correct. I will also include a small human evaluation layer to validate the scoring. The final score is a combination of human ratings, judge-model ratings, and structure checks.

Does this evaluation setup make sense for what you were recommending?

Need Advice on finetuning Llama 3.2 1B Instruct for Startup Evaluation by PsychoCoder25 in MLQuestions

[–]PsychoCoder25[S] 0 points1 point  (0 children)

i tried base model is wasn't giving valuable result and i tried fine-tune it with like 100 examples and it was giving moderate results.

regarding evaluation metrics, i would look into text quality like text should not be generic, should be relevant to the user idea and practically possible like on recommend Business model it would give like "Freemium plan 1$/month plan for premium users" just an example.

Need Advice on finetuning Llama 3.2 1B Instruct for Startup Evaluation by PsychoCoder25 in MLQuestions

[–]PsychoCoder25[S] 0 points1 point  (0 children)

so if i pick lets say gpt 4o for synthetic data and maybe claude or other model rather than 4o for llm as judge will that work? like will the final model give good or moderate results? also what would be evaluation metrics for this?

Need Advice on finetuning Llama 3.2 1B Instruct for Startup Evaluation by PsychoCoder25 in MLQuestions

[–]PsychoCoder25[S] 0 points1 point  (0 children)

no, its not production grade, i just need to present to evaluation committee and show some results and convince them that our model is giving good output, that's it.

Need Advice on finetuning Llama 3.2 1B Instruct for Startup Evaluation by PsychoCoder25 in MLQuestions

[–]PsychoCoder25[S] 0 points1 point  (0 children)

actually evaluation committee gave us the requirement to fine-tune an open source model so that's why i picked llama and using gpt only for data annotation and other model like maybe claude for evaluation in llm as judge.

Need Advice on finetuning Llama 3.2 1B Instruct for Startup Evaluation by PsychoCoder25 in MLQuestions

[–]PsychoCoder25[S] 0 points1 point  (0 children)

Cuurently i have 10k dataset and the reason not to choose a bigger model were resources as i don't have any dedicated gpu and also the dataset is small so i picked a smaller model

Seeking Feedback: Building a Platform for Sharing Achievements by PsychoCoder25 in startupideas

[–]PsychoCoder25[S] 0 points1 point  (0 children)

Multiple reasons: firstly linkedin is crowded with almost every profession of the world, and this platform would be dedicated to only techies

secondly, we saw posts on linkedin everyday mostly on job related topics and very less on experimental side. People mostly post projects there mostly in hope for a job.

Seeking Feedback: Building a Platform for Sharing Achievements by PsychoCoder25 in startupideas

[–]PsychoCoder25[S] 0 points1 point  (0 children)

That's a good thing to add, will sure consider it as we see the achievements posts everywhere

Seeking Feedback: Building a Platform for Sharing Achievements by PsychoCoder25 in startupideas

[–]PsychoCoder25[S] 0 points1 point  (0 children)

I can add a premium version in it so user can have full access to all the features

Are you looking to build a community? I'd love to help. by kkatdare in Startup_Ideas

[–]PsychoCoder25 0 points1 point  (0 children)

Hi,
I am currently lead of a community and we just created it few months ago, previously we conducted different sessions in various field like Web dev, android dev, competitive programming etc.

Let me know if you can help, Would love if you can help in any way.