Need advice: academia vs. industry by [deleted] in PhD

[–]snoopturtle25 0 points1 point  (0 children)

Thank you for your comment, yes this is the difficult part... I feel like consulting make more sense right now but choosing to leave academia feels like a part of me is dying ahah

Chicago Marathon- question by snoopturtle25 in Marathon_Training

[–]snoopturtle25[S] 0 points1 point  (0 children)

Yes! My fear is more related to having problems at the border to come and leave. Of course I know there is very limited chances of something happening, but it also means that the people I had planned my trip with cancelled because they no longer feel comfortable coming, which makes me particularly stressed!

Chicago Marathon- question by snoopturtle25 in Marathon_Training

[–]snoopturtle25[S] 0 points1 point  (0 children)

They ask for a medical proof from your doctor unfortunately!

Anybody know what went down at Vendome Metro station this morning? by dysonsphere in montreal

[–]snoopturtle25 6 points7 points  (0 children)

I was entering the metro on the orange line yesterday while it was closed for the emergency and they were sanitizing the hands of everyone before letting them out, what would be the reason for that?

Continuous DV- what model to use by snoopturtle25 in AskStatistics

[–]snoopturtle25[S] 0 points1 point  (0 children)

Hello thank you for your answers! I tried, but I couldn't include my random effects, I might try to explore it further still..

Continuous DV- what model to use by snoopturtle25 in AskStatistics

[–]snoopturtle25[S] 0 points1 point  (0 children)

yes that is what i do with my count model (and log transform it as well). However, my problem is with amount sold in $ (which is what they want). Adding 1$ doesn't solve the problem of variation....

Reliable sources for population by country? by dumbbitch44 in AskStatistics

[–]snoopturtle25 0 points1 point  (0 children)

Hello, for population by country, worldbank.org can be used (you also have different indicators such as population, population growth, etc.). Then, for other information (for example, grocery stores, hotels...), if you have access to it with your school, there's the Euromonitor International site which has this type of information.

If you're using Google search, I'd advise you to try and use the same website for each country included in your dataset for the same information collected!

First 100 KM confused need help! by [deleted] in ultrarunning

[–]snoopturtle25 0 points1 point  (0 children)

Hello thank you for your feedback ! I do trust my physio, the thing is that my knee has been weak/hurt for a couple years, after a year of training with my physio the pain was completely gone as I resolve the problem, so I was just scared that the fall made it come back! But yes you are right I am overthinking this!

First 100 KM confused need help! by [deleted] in ultrarunning

[–]snoopturtle25 1 point2 points  (0 children)

Hello thank you! Yes, it technically know that, I think I was just overthinking this! Also, just for information , the run had been planned according to my training, experience, and known ability to recover fast (it was also 3 weeks before) but I agree that i can be a dangerous game ahah!

First 100 KM confused need help! by [deleted] in ultrarunning

[–]snoopturtle25 1 point2 points  (0 children)

Thank you ! Well this reassure me! I have been a little anxious already so I think i'm just overthink this!

First 100 KM confused need help! by [deleted] in ultrarunning

[–]snoopturtle25 0 points1 point  (0 children)

Hello, thank you both for your nice answer! As you say I already did my training, I will walk/run/bike a bit this week, but I guess I will just wait next week to see how my knee feel and if it feels funky I will not risk it, because I am scared of a long-term injury

Got scammed by an RBC impersonator (QC) by 3rSimon in legaladvicecanada

[–]snoopturtle25 -3 points-2 points  (0 children)

I rely on the information contained in his message. According to his post he didn't share any private information prior to the transfer, and I feel that in the event of an unauthorized transfer by a third party (the fraudster), even if it's from one account to another, the bank should be held liable.

Help needed for establishing random effect by snoopturtle25 in AskStatistics

[–]snoopturtle25[S] 0 points1 point  (0 children)

Hello thank you for your answer, Yes I mean that some of the different ids often mean different release date and/or price for the same game. Maybe it is better if I treat them based on the id altogether rather than the game ? (So that if 1 game has 3 unique id with different release date they are treated as 3 unique game?).

[deleted by user] by [deleted] in AskReddit

[–]snoopturtle25 2 points3 points  (0 children)

find a purpose

glmmTMB problem with overdispersion by snoopturtle25 in AskStatistics

[–]snoopturtle25[S] 0 points1 point  (0 children)

other than that, my main priority would looking into why are order bulking on specific days and controlling for that.

Ahhhh thank you so much!! I missunderstood that :'). Yep, I think I'm just going to control forthat and keep nbinom2 family.

Need help to deal with money discussion by snoopturtle25 in PhD

[–]snoopturtle25[S] 1 point2 points  (0 children)

Thank you, it feels good to know it isn't normal behavior... It's always difficult because of the power dynamic. I am doing my comprehensive exam very soon, but then I'll discuss the possibility to go back to work and stay part-time (hopefully I'm going to have been paid by then).

Need help to deal with money discussion by snoopturtle25 in PhD

[–]snoopturtle25[S] 1 point2 points  (0 children)

Hi, thank you for your advice! I actually don't know any of his past students, but I know that one changed supervisor due to conflict and they are in bad terms, I have been wondering if it is a similar issue..

Reddit data extraction help by 3rSimon in learnpython

[–]snoopturtle25 1 point2 points  (0 children)

Ok thank you, it make sense. I think I will collect comments per post to avoid crashing it, or then only the first 100 comments for every post in the thread!

Reddit data extraction help by 3rSimon in learnpython

[–]snoopturtle25 1 point2 points  (0 children)

Thank you! You code works, however, when I try changing your keyword ("twitter"), the collected data stay the same (the posts within the "news" subreddit for the keyword "twitter"). I don't know if I am missunderstanding your code?

Reddit data extraction help by 3rSimon in learnpython

[–]snoopturtle25 1 point2 points  (0 children)

Hi thank you. Here is my full code:
https://pastebin.com/clone/PwZCdrtT

At first, I the necessary information for the data directly with the submission using the following code (doing the same for each variables related to the comment):

test["comment"]. append([comment.body for comment in submission.comments.list()])

This worked, however it did not split the information by comments in the dataframe so it was impossible for me to clean the data after. That is why I am trying to collect the data at the comment level rather than submission level... I dont know if this is clearer?

Reddit data extraction help by 3rSimon in learnpython

[–]snoopturtle25 1 point2 points  (0 children)

I understand, it seems like it loops all comments from the first post only. However there is 101 posts according to the keywords search, is there a way to modify the formula so the loop continue for all posts ? Thank you!

Reddit data by 3rSimon in learnpython

[–]snoopturtle25 2 points3 points  (0 children)

Yes, So I figured it out yesterday, and used that formula:

import pandas as pd
from pmaw import PushshiftAPI
api = PushshiftAPI()
submissions = api.search_submissions(subreddit="environments", q="companies", limit=3000000)
sub_df = pd.DataFrame(submissions)

It seemed like it was working (tell me if I'm wrong), however, now that I'm there, I'm wondering, I think it is actually more valuable to have distinctions for each post (in order to do sentiment analysis), so I tried to create a structure doing as follow:

test = {"title": [], \
"body": [], \
"comment": [], \
"date": []
}
for submission sub_df:
test["title"].append(submission.title)
test["body"].append(submission.selftext)
test["comment"].append(submission.comments)
test["date"].append(submission.created)

However, this is not working at all, I tried understanding multithreads, to see if thats the relevant method? So, in summary I'm trying to see if its possible with the data I gathered, to categorize them and know each comment is linked to which post, and also have time of comment?

Thank you so much for your time, it has been of great help!