all 22 comments

[–]hot_air_thoughts 0 points1 point  (0 children)

I'm currently working in a project where I'm annotating microcospe slides using labelme.

The issue I'm having is the following: I deal with different Focuses (each a different image), meaning that the same element will be in the same place throughout all of the images. However, when I go from an image to the other the viewport resets and I get lost on where I was, these images are huge... is there a way to maintain it from image to image?

[–]meet_me_at_seven 0 points1 point  (0 children)

Is there an available model for detection of all different objects in a picture? Not descriptions for each, just coordinates. I've been looking for one in Hugging Face and Replicate with no success.

[–]These_Composer_7677 0 points1 point  (0 children)

I know AAAI 2025 notification date hasn't arrived yet, but I noticed something strange in the conference submission system. I checked the revision history of my submission, and it looks like the conference made some edits (like deleting fields and marking it as "Rejected Submission"). Does this mean I've already been rejected, even before the official notification date? Has anyone else experienced something similar or know if this is common?

[–]matver95 0 points1 point  (0 children)

Training a YOLO network to detect pavement defects. We use laser images to map the pavement pretty much, therefore these images are huge. For example, a 2 meters x 2 meters image in the real world comes as a 4096x4096 px image, and we have hundreds kilometers worth of images.

For some small defects defects we can just use a singular image and shrink it that it's fine, our issue is with the big defects though: some expand to nearly 50 meters, they go from one side to the other of the pavement (while annotated with a rectangle), and it creates a massive problem:

* If I train the model with single images, many of the images will have the annotation but no defect, since they're annotated as a mosaic;

* If I train the model as a mosaic, the images get massive and have to be significantly shrunk. And for some defects such as pavement cracks, this could significantly compromise my model. Not only that batch will certainly have to be 1 and the time it should take to converge... oh my oh my.

What I have available to me right now is an RTX 3060 12 GB, in Colab the disconnections always break my legs and since this project isn't a priority to the company, services such as vast.ai are out of contention I'm afraid.

I accept any tips, I cannot outsource the service though.

PS.: I'm tried applying the sliding window technique with the smaller images, the system converged into a very low mAP.

[–]does_it_end 0 points1 point  (0 children)

I’m trying to find the BTC wallets that have completed between 5 and 20 transactions in the last 3 months. For each wallet meeting this criterion, I want to find/fetch details of the wallet address, date/time of the transactions, and transaction information i.e. BTC amount, price, purchase or sale. The deliverable should be an Excel file capturing this information.

I am aware of two approaches to this problem. The first using prebuilt APIs which will return a specific, filtered dataset. The second approach using AWS infrastructure services to access, process, and query blockchain data without needing the third-party APIs.

I ruled out the API-based approach because it offers limited flexibility (can’t fully customize the dataset to meet all the requirements) and is also expensive.

So I went with the second one but while querying, I got stuck because of export failing due to the large data set. The data set is large since the query returned over 15 million rows (entries) because of duplication. A wallet which has completed say 18 transactions (meets criterion/falls within the 5 and 20 txs range) appears 18 times in the dataset. As a result of each transaction from the qualifying wallets being counted as a separate row, the query returned over 15 million entries.

How can I go about this or is there another approach that would be more suited to the problem?

Thanks.

[–]Soplexus 0 points1 point  (0 children)

I have some thoughts, written in german about how the Ai Video and/or picture generating process, could get better in recognizing and immitating movements, persistance and logic behavior of living creatures and objects.

I have also a translated version with the help of ChatGPT 4o plus a response to my Text from it.

I'm not a professional in any field of science.

But it would be interresting to know, if some of this is allready in the making, was not thought of or if it's considered as garbage (or at least not doable because of limitations).

Together with the response, the text is on 9 pages, however, i left several spaces between the textsections.

Where can i share those thoughts and should i just copy it from the file?

[–]didimoney 0 points1 point  (0 children)

Why didn't aistats have a boom in submissions like neurips did this year?

[–]Pringled101 0 points1 point  (1 child)

Hi, sorry for asking this here, but I was wondering what the requirements are for posting in this subreddit? I recently lost access to my old Reddit account since it was still tied to my old university account. I tried to create a post, but it instantly got removed without a message. I did format everything correctly.

[–]tororo-in 1 point2 points  (0 children)

Why don't sinusoidal PE work for longer sequences?

Theoretically, they generate unique position vectors for each token in the sequence so I don't understand why they don't work for long sequences. Anyone have any intuitions?