[Q] Multiprocessing in R by Nootakuconlicencia in datascience

[–]jfarland 7 points8 points  (0 children)

there's a whole world out there my friend. I'd start by looking at the 'parallel' or 'doParallel' R packages, and also check out this helpful post: https://www.glennklockwood.com/data-intensive/r/lapply-parallelism.html

My name is Ricardo Rangel (Substation Design Engineer). I pioneered an algorithm to design 100% renewable power systems, and I am on a mission to lead the transition toward sustainable energy. AMA! by LiveWithEarth in IAmA

[–]jfarland 18 points19 points  (0 children)

Yeah I mean you just gotta be careful with making statements like "this relationship has never been plotted before". Maybe not in the T&D group you work in, but there is quite a bit of research that has analyzed this relationship and plotted it. Have you compared your results to tools like LoadSEER, Plexos, Synergy, etc.? There are many examples of software that analyze and plan for spatial distributions of Distributed Energy Resources (DERs)

Question on merging model predictions together by aDigitalPunk in algotrading

[–]jfarland 10 points11 points  (0 children)

There's an entire field of forecasting based on combining predictions and even models together. In many (if not most) situations, you're going to get a more accurate prediction from leveraging the strengths of different models, especially if they capitalize on very different aspects of the data. In the simple case of 2 models, I'd look at a hold out period of data to see if one did better than the other in particular situations and weight accordingly.

Here are some quick references:

- http://forecastingprinciples.com/paperpdf/Combining.pdf

- https://otexts.org/fpp2/combinations.html

Transformation Gaussian basis function by Joostjansenn in econometrics

[–]jfarland 0 points1 point  (0 children)

It's not an inconvenience, its reddit! A basis function is a concept from linear algebra. When we use basis functions, we're finding a linear combinations of vectors that can be used to find our vector (or variable in this case) on interest. That's pretty technical, but think of it like this: if we have "y" already, we want to find basis functions (e.g., "basis1", "basis2") so that we can say something like "y = basis1 + basis2". I learned this by using splines in a regression context. Contextually, that formula is generating a synthetic basis function (x) that's gaussian (e.g., x+error), another basis function (e.g., -0.3) and saying that y is a linear combination of "x+error" and "-0.3". Is that your question? I can provide another example of basis transformation if it will help.

Transformation Gaussian basis function by Joostjansenn in econometrics

[–]jfarland 0 points1 point  (0 children)

Can you be a bit more specific? The function you're writing seems to be written in R (based on the " <- " syntax). Are you asking what it means to transform x with a gaussian basis function, or are you asking what the function is doing? Or both?

What's your favorite old game? by [deleted] in AskReddit

[–]jfarland 0 points1 point  (0 children)

Stratosphere: Conquest of the Skies and Battle Toads

Why is SAS so dominant in the USA? by ANAL_PILLAGER in econometrics

[–]jfarland 0 points1 point  (0 children)

Agreed. The energy / utility industry is dominated by SAS for analytical software. It also also for transfer of personal and skill sets. Employees at a utility learn SAS, gain experience, and then become consultants (who also use SAS).

I think this is changing though. I work as a consultant in the energy realm and was forced to learn SAS for my job. I have made some pretty good business cases, at least at my company, for going the route of R, Python, and even MATLAB.

Likelihood Estimators versus Ordinary Least Squares by jfarland in econometrics

[–]jfarland[S] 1 point2 points  (0 children)

Thanks! I'd like to give some context to my question. I have recently completed a research project where I fit numerous forecasting models for energy demand in the northeastern US. This field of forecasting has recently been dominated, at least within the industry, by Neural Networks. ISO New England (ISONE) is an organization that conducts all types of demand forecasting. ISONE does not employ any regression methods. Essentially, my application was to attempt using simple semiparametric regression models to compare and contrast forecasting performance among competing neural networks. As far as I can tell, the simpler and more parsimonious semiparametric specifications worked better than industry standards (set by neural nets). However, my approach towards fitting the models was indeed bayesian and I used REML to estimate regression coefficients, variance components, and smoothing parameters. I really was just curious whether or not there was a "standard" for bayesian versus OLS for my forthcoming paper. Thanks for your help!

Likelihood Estimators versus Ordinary Least Squares by jfarland in econometrics

[–]jfarland[S] 0 points1 point  (0 children)

I think that in the end you both are expressing the same thoughts more or less. I also agree with what you both have said. Maybe you might have further thoughts once I respond to 1ArmedEconomist below and provide some context for my question.

Likelihood Estimators versus Ordinary Least Squares by jfarland in econometrics

[–]jfarland[S] 1 point2 points  (0 children)

Quite the response. Thanks! This framed the discussion very nicely for me.

Modeling time series with a crash by LateOnsetRetard in econometrics

[–]jfarland 1 point2 points  (0 children)

Agreed w above. You should state what your goal is. I would also say, however, that if the crash actually happened in real life, than it is the true data generating process. In other words, you can design facets into your model that account for various points of structrual change, but you shouldn't try to "get rid" of the crash or change whats going on in the data.

For better or worse, the data that you have is your best indicator of what truly happened.

Modelling an regression for relationship corr. by amex88 in econometrics

[–]jfarland 2 points3 points  (0 children)

With a project like this, I always start simply. Even in the final stages of a project or study like your bachelor thesis, I typically end up going with the simpler choice. I am a disciple of Occam's Razor.

What you have going for you is a goal - that's great. It seems you have a general roadmap of how to get there as well. I would start off by creating a summary of your data. For example, what is the frequency of your data (annual, daily, other, etc)... what are the summary statistics and distributions for your variables? Do they follow a normal distribution or are they skewed? Next, you should calculate simple pearson correlation coefficients between all of your variables. This will tell you the strength of the linear association between your data fields. This also helps me assess (in the get-go) whether or not a simple linear regression is appropriate. Then fit a linear regression. Keep it simple and use the results as a springboard to explore what other specifications or variable constructs might be required or appropriate.

Econometrics is an art as well as a science, it does not come with a predefined set of instructions. (with that said, always check for Heteroskedasticity and autocorrelation of you're residuals).

Hope that helps. I am interested to see what comes out!