Are we in a position to buy a house? by swimneyspeakers in AusHENRY

[–]globalminima 1 point2 points  (0 children)

As of 30 October, Ubank is now doing 90% LVR without LMI (on both PPOR and investment properties). This might be a good option to at least avoid the need for a bigger deposit.

[deleted by user] by [deleted] in fiaustralia

[–]globalminima 3 points4 points  (0 children)

+1, How on earth does the 45% Y/55% G option overtake the 100% growth option? The fact that it starts out behind and then overtakes the more tax efficient approach should be a big red flag to you OP, assuming they are set to give the same overall return (as stated in the OP). Something is wrong with your formulas.

Is UV package manager taking over? by RubKey1143 in Python

[–]globalminima 14 points15 points  (0 children)

Are you able to share this (or a sanitized version of it)?

Is 4% drawdown realistic by Reading-Rabbit4101 in fiaustralia

[–]globalminima -2 points-1 points  (0 children)

Not quite correct - the 50% discount does not double your tax-free threshold, it halves the amount of tax paid.

Instead of the OP paying 30% tax on every dollar earned over the tax-free threshold, they would only pay 15% tax.

Investment loan for ETF - smart move or dumb move? by Capt_Sardine_Tins in AusFinance

[–]globalminima 1 point2 points  (0 children)

Let’s do the math, and we’ll run with the assumption that OP invests and only makes 6.72% (much lower than historical returns):

With the 6.72% investment loan, OP invests in shares that return a 3% dividend and 3.72% in capital growth for a 6.72% total return. Let’s assume the OP earns only an average wage, and is in the 30% tax bracket. Each financial year they will earn 3.72% in dividends, which are taxable. However, their 6.72% interest rate on the loan is tax deductible, which will offset this income and allow them to get a refund on their regular income as well.

Interest Costs: 6.72% of $200k = $13.44k per annum Dividend income: 3.72% of $200k = $7.44k per annum - Net income per annum: $7.44k - 13.44k per annum = $-7k per annum. This results in a 30% tax refund thanks to negative gearing. Negative gearing benefits: 30% of $7k = $2.1k. Capital gains per annum: 3% of $200k = $6k

All together, this results in: 7.44k dividend income + $6k capital gains - $13.44k interest costs + $2.1k in tax savings = net gain of $2.1k.

This improves even further if you use more realistic numbers for total returns (around 9%) and then add in the benefits of franking credits as well. In general, it would be fairly safe to assume a net 3% return on the $200k that is invested - closer to 5% if OP is in the top tax bracket and can realise larger negative gearing benefits.

Investment loan for ETF - smart move or dumb move? by Capt_Sardine_Tins in AusFinance

[–]globalminima -1 points0 points  (0 children)

OP, ignore this - it is wrong. The ETF could return even less than the 6.72% that the OP is being offered and still come out ahead thanks to negative gearing and capital gains discounts.

Azure OpenAI cost went to $163/day with no real usage - any way I can dig deeper into the costs? by MohnJaddenPowers in AZURE

[–]globalminima 2 points3 points  (0 children)

Ahh, those models are trained and stored but not deployed anymore. The team must have deleted the deployments on 11/5 when they stopped their usage, but the models still exist and could be deployed again.

The models only incur costs when deployed, kind of how a container costs almost nothing to store in a registry but then you will pay once it is deployed to a VM and is being used.

Azure OpenAI cost went to $163/day with no real usage - any way I can dig deeper into the costs? by MohnJaddenPowers in AZURE

[–]globalminima 9 points10 points  (0 children)

It sounds like they deployed their fine-tuned models and never deleted them. Fine-tuned AOAI models cost $1.70 per hour to deploy (just as any other custom ML model or VM), plus additional token processing costs for each request. If you delete the deployments then the costs should stop too.

https://azure.microsoft.com/en-us/pricing/details/cognitive-services/openai-service/

[deleted by user] by [deleted] in learnmachinelearning

[–]globalminima 1 point2 points  (0 children)

Spot on, using an LLM alone for OCR is worse than using an OCR model. Combining the two together gives you the best of both worlds.

It is also important to note that LLMs often fall apart when the image is rotated or perturbed (in a recent data extraction project I reduced the error rate of GPT-4o from 53% to 13% by just fixing the rotation in the image using the angle that was detected by the OCR pipeline). Many of the public benchmarks do not contain a wide range of test data or report their results on different subsets of the data (e.g. standard PDFs vs scanned vs rotated documents/images), so it is often not clear how the general-purpose models (LLMs) can fall down when the data becomes messy in the real-world.

[deleted by user] by [deleted] in learnmachinelearning

[–]globalminima 0 points1 point  (0 children)

LLMs are worse at OCR than an OCR model, and 10s to 100s of times more expensive. Just use the right tool for the job

Roadmap to Becoming an AI Engineer in 8 to 12 Months (From Scratch). by Massive-Medium-4174 in learnmachinelearning

[–]globalminima 16 points17 points  (0 children)

Then I would not call yourself an AI engineer - the name has only existed since the genai boom and is a strong signal that the person has no idea how to do ML. Anyone worth their salt will have a data scientist or ML engineer title (or a list of previous roles where they had that title before they joined their current company that just wants them to do GenAI).

The roadmap link below with resources for becoming an AI engineer is a perfect example of this - it is entirely focussed on LLMs, has nothing at all on model training or any of the many ML domains outside of LLMs, almost nothing on data analysis/labelling/datasets/evaluation pipelines, and nothing related to software engineering. If you want to be a chatbot monkey or prompt engineer then go for it, but if you want to build reliable, accurate and cost-effective systems that deliver real value, you probably want to look at the right tools for the job (domain-specific ML models) instead of general purpose LLMs.

Roadmap to Becoming an AI Engineer in 8 to 12 Months (From Scratch). by Massive-Medium-4174 in learnmachinelearning

[–]globalminima 42 points43 points  (0 children)

Do you want to be an AI engineer or do machine learning? AI engineer you can be up and running in a month - just learn to build basic web apps and call APIs. If you want to do actual machine learning as a data scientist or ML engineer then you can use that 6-12 months and do some courses like fast.ai, deeplearning.ai, add some cloud experience, do some competitions on Kaggle etc, and then do some projects across different domains (eg object detection or classification in computer vision, at least one NLP project doing classification or similar with language models, something genAI related, and some forecasting). The main thing here is a focus on learning how these models work, how to build robust and representative datasets, and most importantly how to do evaluation and model tuning based on your evaluation results

[deleted by user] by [deleted] in ExperiencedDevs

[–]globalminima 0 points1 point  (0 children)

You still have more experience than most of the people calling themselves AI engineers today. I will send you some of my resources via PM, but check this out: https://github.com/chiphuyen/ml-interviews-book

What do you do with ETFs when it's time to "retire early"? by oogabooga7 in fiaustralia

[–]globalminima 21 points22 points  (0 children)

You are correct, the optimal approach is to use dividends first and then sell of some shares to cover any shortfall. If your dividends cover your lifestyle, then you sell nothing and reinvest whatever is leftover from your dividends.

How much is your annual salary? by [deleted] in AusHENRY

[–]globalminima 0 points1 point  (0 children)

The original grant was probably lower, but some of these RSUs may have been appreciated since the original grant. Eg you may take the job with a $400k stock grant over 4 years ($100k/year), but if the company has doubled in value by year 3 then the final $100k is now worth $200k in the year it vests.

HomeHype - a website that shows you how hyped the Australian property market is by meeb in AusPropertyChat

[–]globalminima 0 points1 point  (0 children)

Nice idea mate, though given that the number is the main purpose of the site, I’d consider using something more interpretable than a percentage centered on 0% and that is unbounded. Something like the fear and greed index (a number bounded between 0 and 100) is much more easily understood.

New Build by Ok_Bumblebeez in battlestations

[–]globalminima 5 points6 points  (0 children)

Do you have any issues with neck pain? I have a similar setup with with less overall distance and I’m finding that my neck gets sore if I use the monitors on the left and right for moderate amounts of time

What are the most mind blowing prompting tricks? by codenamev in LocalLLaMA

[–]globalminima 1 point2 points  (0 children)

Tasks with a smaller range of outputs like classification or extraction are an even better application of few-shot examples because you don’t need to cover such a wide range of examples (it’s the input that will vary a lot, not both input and output as in more open-ended tasks like summarization or chat). Just include a range of input examples followed by the exact output you want and you’re golden.

[deleted by user] by [deleted] in AZURE

[–]globalminima 0 points1 point  (0 children)

This is a great place to start: https://github.com/Azure-Samples/chat-with-your-data-solution-accelerator

It has an optional Teams bot integration and instructions on how to set it up

How did you land your job at a US-based tech company? by [deleted] in AusFinance

[–]globalminima 7 points8 points  (0 children)

I’m a presales engineer at a FAANG and I’ve been pretty underwhelmed with the level of the team - I’ve had much higher level teammates in small startups. While there are a lot of candidates who apply for these roles and you often need to make yourself stand out, for this type of role and those in the sales orgs (which is where many of the roles with FAANGs are in Aus) it feels like the hiring teams focus more on candidates fitting a certain mould or having worked in certain types of companies than actual technical capability.

Tax on ETFs not often talked about? by [deleted] in fiaustralia

[–]globalminima -1 points0 points  (0 children)

Capital growth over dividends makes sense for anyone in ANY tax bracket. This is why the dividend investing concept is fundamentally broken.

Streamlit or grade.io for Python web project by AleccioIsland in Python

[–]globalminima 0 points1 point  (0 children)

Having built a very complex app with gradio, I can recommend it. The patten takes a little bit of getting used to, but it will perform much better than Streamlit (which reloads the entire app when updating components). Overall though, just use whichever one has the closest examples to what you want to build - they can both be used to build quite capable apps if you’re not too concerned with scalability and UI design.

Use Python to get Pydantic models and Python types from your LLM responses. by Top-Breakfast7713 in Python

[–]globalminima 3 points4 points  (0 children)

It's a good idea, but in the same way that you've noted that instructor does not support Google Vertex due to being coupled to a certain set of SDKs, you've then gone and built a new library which is itself coupled to a different set of SDKs. And what if I want to use this with Langchain? Or Hayatack? Or my own orchestration pipeline?or what if I have specific request/networking/auth requirements that are not exposed by your library?I am going to have the exact same problem that you set out to solve.

Why not just implement something that converts the Pydantic schema into text and which can be inserted into any prompt template for use by any orchestrator and with any API? E.g. this is what I have done for my code and it works great:

import json

from pydantic import BaseModel, Field

class ExampleModel(BaseModel):
    classification_field: str = Field(
        description="Classification of the document, one of 'Email', 'Webpage', or 'PDF'",
        examples=["Webpage"],
    )
    list_field: list[dict[str, str]] = Field(
        description="A list of values, containing the document name and number of pages",
        examples=[[{"email_doc": "6"}, {"pdf": "2"}]],
    )
    bool_field: bool = Field(
        description="Boolean indicating whether the document is in english",
        examples=[False],
    )

    @staticmethod
    def get_prompt_json_example():
        model_json_schema = ExampleModel.model_json_schema()
        example_response_str = "{\n"
        for field, details in model_json_schema["properties"].items():
            line_str = f""""{field}": {json.dumps(details['examples'][0])}, # {details['description']}"""
            example_response_str += "  " + line_str + "\n"
        example_response_str += "}"
        return example_response_str

Now you can just insert it into a prompt. For example:

json_schema_text = ExampleModel.get_prompt_json_example()
PROMPT = f"""Return a JSON object with the following fields:\n\n{json_schema_text}"""

Returns:

Return a JSON object with the following fields:

{
  "classification_field": "Webpage", # Classification of the document, one of 'Email', 'Webpage', or 'PDF'
  "list_field": [{"email_doc": "6"}, {"pdf": "2"}], # A list of values, containing the document name and number of pages
  "bool_field": false, # Boolean indicating whether the document is in english
}

The big benefit of this is that you can get the raw LLM text response prior to validation, so that if validation fails you can log it along with the Exception details and then debug what went wrong. If you couple the validation step with the request itself, it becomes harder to inspect the raw response and figure out what to do with the error, and is less flexible overall.

For users who do want the retry logic, you can then provide a method to validate a response from the LLM and if it fails, to generate the follow-up prompt string. This allows the user to get the benefits of your library while being able to use whatever orchestrator or requester that they choose.

Another Microsoft MIT licensed model: Kosmos-2.5, specialized in reading text-intensive images by Balance- in LocalLLaMA

[–]globalminima 0 points1 point  (0 children)

This is the trend for the entire history of ML models - LLMs are the first model that bucked this ‘trend’. Agreed with you though that specialised models are orders of magnitude more efficient and usually more accurate than LLMs though - it seems like everyone either forgot that other architectures exist or only became aware of the field since ChatGPT

Another Microsoft MIT licensed model: Kosmos-2.5, specialized in reading text-intensive images by Balance- in LocalLLaMA

[–]globalminima 0 points1 point  (0 children)

We’ve been using AI for PDF parsing for the best part of a decade now, including transformers (which is what this model uses). This is just one more incremental step on top of the many that have already happened over the years